Hotnews

lesswrong

lesswrong 25h ago 19°

On The Adolescence of Technology

Published on January 30, 2026 12:50 PM GMTAnthropic CEO Dario Amodei is back with another extended essay, The Adolescence of Technology.
lesswrong 46h ago 14°

LLM Alignment, ethical and mathematical realism, and the most important actions in davidad's understanding

Published on January 29, 2026 3:48 PM GMTIntroduction to davidad and today's topicstutor valsLessWrong prides itself for an ethos of "say it how you think it" (see "A case for courage when speaking of AI danger").
lesswrong 41h ago 13°

The Wolves Are All Gone

Published on January 29, 2026 8:24 PM GMTChapter 1: The Prompt
lesswrong 48h ago 12°

The third option in alignment

Published on January 29, 2026 2:20 PM GMTUsually the doom conversation is binary. Either the AI will try to take over the world for his own reward function as is discussed in all the alignment problem debates OR a malicious dictator or bad actor will leverage the technology's power to take over the world, create 1984's style of government or many other doom scenarios such as nukes and bio-weapons.
lesswrong 20h ago 12°

Addressing Objections to the Intelligence Explosion

Published on January 30, 2026 6:21 PM GMT1 IntroductionCrosspost of this blog post. My guess is that there will soon be an intelligence explosion.I think the world will witness extremely rapid economic and technological advancement driven by AI progress. I’d put about 60% odds on the kind of growth depicted variously in AI 2027 and Preparing For The Intelligence Explosion (PREPIE), with GDP growth rates well above 10%, and maybe above 100%.
lesswrong 39h ago 11°

How to Hire a Team

Published on January 29, 2026 10:39 PM GMTA low-effort guide I dashed off in less than an hour, because I got riled up. Try not to hire a team. Try pretty hard at this.Try to find a more efficient way to solve your problem that requires less labor – a smaller-footprint solution.Try to hire contractors to do specific parts that they’re really good at, and who have a well-defined interface.
lesswrong 41h ago 10°

Problems with "The Possessed Machines"

Published on January 29, 2026 9:00 PM GMTSo, The Possessed Machines. There's been some discussion already. It is a valuable piece -- it has certainly provoked some thought in me! -- but it has some major flaws. It (sneakily!) dismisses specific arguments about AI existential risk and broad swaths of discourse altogether without actually arguing against them.
lesswrong 45h ago 10°

Are We in a Continual Learning Overhang?

Published on January 29, 2026 5:09 PM GMTSummary: Current AI systems possess superhuman memory in two forms, parametric knowledge from training and context windows holding hundreds of pages, yet no pathway connects them. Everything learned in-context vanishes when the conversation ends, a computational form of anterograde amnesia. Recent research suggests weight-based continual learning may be closer than commonly assumed.
lesswrong 45h ago

Bentham’s Bulldog is wrong about AI risk

Published on January 29, 2026 4:33 PM GMT(...but also gets the most important part right.)Bentham’s Bulldog (BB), a prominent EA/philosophy blogger, recently reviewed If Anyone Builds It, Everyone Dies. In my eyes a review is good if it uses sound reasoning and encourages deep thinking on important topics, regardless of whether I agree with the bottom line. Bentham’s Bulldog definitely encourages deep, thoughtful engagement on things that matter.
lesswrong 17h ago

36,000 AI Agents Are Now Speedrunning Civilization

Published on January 30, 2026 9:21 PM GMTPeople's Clawdbots now have their own AI-only Reddit-like Social Media called Moltbook and they went from 1 agent to 36k+ agents in 72 hours.As Karpathy puts it:What's currently going on at @moltbook is genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently. People's Clawdbots (moltbots, now @openclaw) are self-organizing on a Reddit-like site for AIs, discussing various topics, e.g.
lesswrong 16h ago

Monitoring benchmark for AI control

Published on January 30, 2026 9:13 PM GMTMonitoring benchmark/Semi-automated red-teaming for AI controlWe are a team of control researchers supported by CG’s Technical AI safety grant. We are now halfway through our project and would like to get feedback on the following contributions.
lesswrong 45h ago

Disempowerment patterns in real-world AI usage

Published on January 29, 2026 4:36 PM GMT[W]e’re publishing a new paper that presents the first large-scale analysis of potentially disempowering patterns in real-world conversations with AI.
lesswrong 41h ago

Better evals are not enough to combat eval awareness

Published on January 29, 2026 8:42 PM GMTIntroThe eval awareness is a pressing problem, and in this post I advocate that just making benchmarks more realistic won't solve the evaluation awareness problem alone. Better evals must be coupled with other measures, like change in the way models are trained.Safety evaluations are supposed to predict how models will behave in deployment.
lesswrong 14h ago

Prediction: Recursively Self-improving AI for 2033

Published on January 30, 2026 11:53 PM GMTContext:One way to measure how good LLMs are which is gaining traction and validity is the following:Let T(t) be the time it takes for a human to the task t.
lesswrong 17h ago

The Matchless Match

Published on January 30, 2026 9:18 PM GMTtl;dr I wanted to consult the internet for good triple and quadruple entendres. unable to find a good existing compilation, so I wrote my own!
lesswrong 14h ago

Forecast: Recursively Self-improving AI for 2033

Published on January 30, 2026 11:53 PM GMTContext:One way to measure how good LLMs are which is gaining traction and validity is the following:Let T(t) be the time it takes for a human to the task t.
1 2 3