LLMs are stuck in Plato’s cave — LessWrong
AI

LLMs are stuck in Plato’s cave — LessWrong

Epistemic status: This is an intuition pump for why making LLMs multimodal is helpful. I am ~70% confident that my later claims which build on this are at least somewhat correct. This article was created over an unclear number of hours (20-50?) and red teamed by gpt-4o. I recently went to a festival at a […]

Binary and analog variation of synapses between cortical pyramidal neurons — LessWrong
AI

Binary and analog variation of synapses between cortical pyramidal neurons — LessWrong

Join Us for the Memory Decoding Journal Club!  A collaboration of the Carboncopies Foundation and BPF Aspirational Neuroscience This time, we’re diving into a groundbreaking paper:“Binary and analog variation of synapses between cortical pyramidal neurons” Authors: Sven Dorkenwald, Nicholas L Turner, Thomas Macrina, Kisuk Lee, Ran Lu, Jingpeng Wu, Agnes L Bodor, Adam A Bleckert, Derrick Brittain,

Stop and check! The parable of the prince and the dog — LessWrong
AI

Stop and check! The parable of the prince and the dog — LessWrong

This post is a response to John Wentsworth’s recent post on Generalized Hangriness specifically as it applies outrage, an emotion that is especially likely to make false claims. I expect that some readers will find it obvious, but I hope others will find it useful to have the concept laid out clearly. My aim is

Stop and check! The parable of the prince and the dog — LessWrong
AI

Take Precautionary Measures Against Superhuman AI Persuasion — LessWrong

Please consider minimizing direct use of AI chatbots (and other text-based AI) in the near-term future, if you can. The reason is very simple: your sanity may be at stake. Commercially available AI already appears capable of inducing psychosis in an unknown percentage of users. This may not require superhuman abilities: It’s fully possible that

Vitalik’s Response to AI 2027 — LessWrong
AI

Vitalik’s Response to AI 2027 — LessWrong

Daniel notes: This is a linkpost for Vitalik’s post. I’ve copied the text below so that I can mark it up with comments. … Special thanks to Balvi volunteers for feedback and review In April this year, Daniel Kokotajlo, Scott Alexander and others released what they describe as “a scenario that represents our best guess

Cybersecurity’s global alarm system is breaking down
AI

Cybersecurity’s global alarm system is breaking down

Cybersecurity practitioners have since flooded Discord channels and LinkedIn feeds with emergency posts and memes of “NVD” and “CVE” engraved on tombstones. Unpatched vulnerabilities are the second most common way cyberattackers break in, and they have led to fatal hospital outages and critical infrastructure failures. In a social media post, Jen Easterly, a US cybersecurity

Prompt Perturbations Reveal Human-Like Biases in LLM Survey Responses
AI

Prompt Perturbations Reveal Human-Like Biases in LLM Survey Responses

arXiv:2507.07188v1 Announce Type: new Abstract: Large Language Models (LLMs) are increasingly used as proxies for human subjects in social science surveys, but their reliability and susceptibility to known response biases are poorly understood. This paper investigates the response robustness of LLMs in normative survey contexts — we test nine diverse LLMs on questions from the

Prompt Perturbations Reveal Human-Like Biases in LLM Survey Responses
AI

Simulated Benchmark with Structured and Unstructured Medical Records

[Submitted on 13 Sep 2024 (v1), last revised 10 Jul 2025 (this version, v3)] View a PDF of the paper titled SimSUM: Simulated Benchmark with Structured and Unstructured Medical Records, by Paloma Rabaey and 2 other authors View PDF HTML (experimental) Abstract:Clinical information extraction, which involves structuring clinical concepts from unstructured medical text, remains a

Scroll to Top