The Download: how your data is being used to train AI, and why chatbots aren’t doctors
AI

The Download: how your data is being used to train AI, and why chatbots aren’t doctors

Millions of images of passports, credit cards, birth certificates, and other documents containing personally identifiable information are likely included in one of the biggest open-source AI training sets, new research has found. Thousands of images—including identifiable faces—were found in a small subset of DataComp CommonPool, a major AI training set for image generation scraped from […]

Monthly Roundup #32: July 2025 — LessWrong
AI

Monthly Roundup #32: July 2025 — LessWrong

Welcome to the monthly roundup of things that don’t fit into other categories and don’t rise to the level of their own posts. Bad News When people tell you who they are, believe them (with obvious exceptions). In particular, if they explicitly describe themselves as evil, or demonic, or uses other similar terms, definitely believe

Building a Smart Python-to-R Code Converter with Gemini AI-Powered Validation and Feedback
AI

Building a Smart Python-to-R Code Converter with Gemini AI-Powered Validation and Feedback

class EnhancedPythonToRConverter: “”” Enhanced Python to R converter with Gemini AI validation “”” def __init__(self, gemini_api_key: str = None): self.validator = GeminiValidator(gemini_api_key) self.import_mappings = { ‘pandas’: ‘library(dplyr)\nlibrary(tidyr)\nlibrary(readr)’, ‘numpy’: ‘library(base)’, ‘matplotlib.pyplot’: ‘library(ggplot2)’, ‘seaborn’: ‘library(ggplot2)\nlibrary(RColorBrewer)’, ‘scipy.stats’: ‘library(stats)’, ‘sklearn’: ‘library(caret)\nlibrary(randomForest)\nlibrary(e1071)’, ‘statsmodels’: ‘library(stats)\nlibrary(lmtest)’, ‘plotly’: ‘library(plotly)’, } self.function_mappings = { ‘pd.DataFrame’: ‘data.frame’, ‘pd.read_csv’: ‘read.csv’, ‘pd.read_excel’: ‘read_excel’, ‘df.head’: ‘head’, ‘df.tail’: ‘tail’,

Automated Face Blurring and Human Movement Kinematics Extraction from Videos Recorded in Clinical Settings
AI

Automated Face Blurring and Human Movement Kinematics Extraction from Videos Recorded in Clinical Settings

[Submitted on 21 Feb 2024 (v1), last revised 18 Jul 2025 (this version, v2)] View a PDF of the paper titled SecurePose: Automated Face Blurring and Human Movement Kinematics Extraction from Videos Recorded in Clinical Settings, by Rishabh Bajpai and Bhooma Aravamuthan View PDF Abstract:Movement disorder diagnosis often relies on expert evaluation of patient videos,

Automated Face Blurring and Human Movement Kinematics Extraction from Videos Recorded in Clinical Settings
AI

[2409.04617] Sparse Rewards Can Self-Train Dialogue Agents

[Submitted on 6 Sep 2024 (v1), last revised 18 Jul 2025 (this version, v3)] View a PDF of the paper titled Sparse Rewards Can Self-Train Dialogue Agents, by Barrett Martin Lattimer and 3 other authors View PDF Abstract:Recent advancements in state-of-the-art (SOTA) Large Language Model (LLM) agents, especially in multi-turn dialogue tasks, have been primarily

Automated Face Blurring and Human Movement Kinematics Extraction from Videos Recorded in Clinical Settings
AI

Autonomous High-Quality Image Editing Triplet Mining

[Submitted on 18 Jul 2025] View a PDF of the paper titled NoHumansRequired: Autonomous High-Quality Image Editing Triplet Mining, by Maksim Kuprashevich and Grigorii Alekseenko and Irina Tolstykh and Georgii Fedorov and Bulat Suleimanov and Vladimir Dokholyan and Aleksandr Gordeev View PDF Abstract:Recent advances in generative modeling enable image editing assistants that follow natural language

AI Safety course intro blog
AI

AI Safety course intro blog

Published on July 21, 2025 2:35 AM GMT This is a linkpost for the intro course for CS 2881: AI Safety. It is mostly intended for Harvard/MIT students considering taking the course but could be of interest to others. Discuss

What Eliezer got wrong about evolution — LessWrong
AI

What Eliezer got wrong about evolution — LessWrong

This post is for deconfusing:  Ⅰ. what is meant with AI and evolution. Ⅱ. how evolution actually works.Ⅲ. the stability of AI goals.Ⅳ. the controllability of AI. Along the way, I address some common conceptions of each in the alignment community, as described well but mistakenly by Eliezer Yudkowsky.  Ⅰ. Definitions and distinctions By far the

Scroll to Top