Natural Intelligence is Overhyped — LessWrong
Published on January 21, 2025 6:09 PM GMTLike this piece? It's cross-posted from by blog: https://collisteru.net/writing/This...
14+ AI Safety Advisors You Can Speak to – New AISafety.com Resource — LessWrong
Published on January 21, 2025 5:34 PM GMTGetting personalised advice from a real human can help...
[Linkpost] Why AI Safety Camp struggles with fundraising (FBB #2) — LessWrong
Published on January 21, 2025 5:27 PM GMTCrossposted on The Field Building Blog and the EA...
The Manhattan Trap: Why a Race to Artificial Superintelligence is Self-Defeating — LessWrong
Published on January 21, 2025 4:57 PM GMTDiscuss
Links and short notes, 2025-01-20 — LessWrong
Published on January 21, 2025 4:10 PM GMTMuch of this content originated on social media. To follow...
The Case Against AI Control Research — LessWrong
Published on January 21, 2025 4:03 PM GMTThe AI Control Agenda, in its own words:… we argue...
Will AI Resilience protect Developing Nations? — LessWrong
Published on January 21, 2025 3:31 PM GMTPosition Piece: Most of the developing world lacks the...
Sleep, Diet, Exercise and GLP-1 Drugs — LessWrong
Published on January 21, 2025 12:20 PM GMTAs always, some people need practical advice, and we...
We don't want to post again "This might be the last AI Safety Camp" — LessWrong
Published on January 21, 2025 12:03 PM GMTWe still need more funding to be able to...
On Responsibility — LessWrong
Published on January 21, 2025 10:47 AM GMTMy view on the concept of responsibility has shifted...
The Gentle Romance — LessWrong
Published on January 19, 2025 6:29 PM GMTA story I wrote about living through the transition...
Is theory good or bad for AI safety? — LessWrong
Published on January 19, 2025 10:32 AM GMTWe choose to go to the moon in this...
What's the Right Way to think about Information Theoretic quantities in Neural Networks? — LessWrong
Published on January 19, 2025 8:04 AM GMTTl;dr, Neural networks are deterministic and sometimes even reversible,...
Per Tribalismum ad Astra — LessWrong
Published on January 19, 2025 6:50 AM GMTCapitalism is powered by greed. People want to make...
Shut Up and Calculate: Gambling, Divination, and the Abacus as Tantra — LessWrong
Published on January 19, 2025 3:03 AM GMTTHERE ARE LAKES at the bottom of the ocean....
Five Recent AI Tutoring Studies — LessWrong
Published on January 19, 2025 3:53 AM GMTLast week some results were released from a 6-week...
be the person that makes the meeting productive — LessWrong
Published on January 18, 2025 10:32 PM GMTHow many times have you been in a meeting...
Well-being in the mind, and its implications for utilitarianism — LessWrong
Published on January 18, 2025 3:32 PM GMTWhen learning about classic utilitarianism (approximately, the quest to...
How likely is AGI to force us all to be happy forever? (much like in the Three Worlds Collide novel) — LessWrong
Published on January 18, 2025 3:39 PM GMTHi, everyone. I'm not sure if my post is...
Beards and Masks? — LessWrong
Published on January 18, 2025 4:00 PM GMT In general, you're not supposed to wear a...
[Exercise] Two Examples of Noticing Confusion — LessWrong
Published on January 18, 2025 3:29 PM GMTConfusion is a felt sense; a bodily sensation you...
Scaling Wargaming for Global Catastrophic Risks with AI — LessWrong
Published on January 18, 2025 3:10 PM GMTWe’re developing an AI-enabled wargaming-tool, grim, to significantly scale...
Alignment ideas — LessWrong
Published on January 18, 2025 12:43 PM GMTepistemic status: I know next to nothing about evolution,...
Don’t ignore bad vibes you get from people — LessWrong
Published on January 18, 2025 9:20 AM GMTI think a lot of people have heard so...
Renormalization Redux: QFT Techniques for AI Interpretability — LessWrong
Published on January 18, 2025 3:54 AM GMTIntroduction: Why QFT?In a previous post, Lauren offered a...
Your AI Safety focus is downstream of your AGI timeline — LessWrong
Published on January 17, 2025 9:24 PM GMTCross-posted from SubstackFeeling intellectually understimulated, I've begun working my...
What are the plans for solving the inner alignment problem? — LessWrong
Published on January 17, 2025 9:45 PM GMTInner Alignment is the problem of ensuring mesa-optimizers (i.e....
Experts' AI timelines are longer than you have been told? — LessWrong
Published on January 16, 2025 6:03 PM GMTThis is a linkpost for How should we analyse...
Numberwang: LLMs Doing Autonomous Research, and a Call for Input — LessWrong
Published on January 16, 2025 5:20 PM GMTSummaryCan LLMs science? The answer to this question can...
Topological Debate Framework — LessWrong
Published on January 16, 2025 5:19 PM GMTI would like to thank Professor Vincent Conitzer, Caspar...
AI #99: Farewell to Biden — LessWrong
Published on January 16, 2025 2:20 PM GMTThe fun, as it were, is presumably about to...
Deceptive Alignment and Homuncularity — LessWrong
Published on January 16, 2025 1:55 PM GMTNB this dialogue occurred at the very end of...
Introducing the WeirdML Benchmark — LessWrong
Published on January 16, 2025 11:38 AM GMTWeirdML websiteRelated posts:How good are LLMs at doing ML...
Replicators, Gods and Buddhist Cosmology — LessWrong
Published on January 16, 2025 10:51 AM GMTFrom the earliest days of evolutionary thinking, we’ve used...
Quantum without complication — LessWrong
Published on January 16, 2025 8:53 AM GMTLearning quantum mechanics involves two things:learning the fundamentals, the...
Permanents: much more than you wanted to know — LessWrong
Published on January 16, 2025 8:04 AM GMTToday's "nanowrimo" post is a fun longform introduction to permanents...
Gaming TruthfulQA: Simple Heuristics Exposed Dataset Weaknesses — LessWrong
Published on January 16, 2025 2:14 AM GMT(Explanation. Also I have no reason to think they...
Applications Open for the Cooperative AI Summer School 2025! — LessWrong
Published on January 15, 2025 6:16 PM GMTApplications are now open for the Cooperative AI Summer School,...
List of AI safety papers from companies, 2023–2024 — LessWrong
Published on January 15, 2025 6:00 PM GMTI'm collecting (x-risk-relevant) safety research from frontier AI companies...
AI Alignment Meme Viruses — LessWrong
Published on January 15, 2025 3:55 PM GMTSome fraction of the time, LLMs naturally go on...
Looking for humanness in the world wide social — LessWrong
Published on January 15, 2025 2:50 PM GMTSocial networks have shaped me since a young age....
On the OpenAI Economic Blueprint — LessWrong
Published on January 15, 2025 2:30 PM GMTTable of Contents Man With a Plan. Oh the...
A problem shared by many different alignment targets — LessWrong
Published on January 15, 2025 2:22 PM GMTThe first section describes problems with a few different...
LLMs for language learning — LessWrong
Published on January 15, 2025 2:08 PM GMTMy current outlook on LLMs is that they are...
Feature request: comment bookmarks — LessWrong
Published on January 15, 2025 6:45 AM GMTSometimes I see a comment I'd like to bookmark,...
How do fictional stories illustrate AI misalignment? — LessWrong
Published on January 15, 2025 6:11 AM GMTThis is an article in the featured articles series...
We probably won't just play status games with each other after AGI — LessWrong
Published on January 15, 2025 4:56 AM GMTThere is a view I’ve encountered somewhat often,[1] which can...
Progress links and short notes, 2025-01-13 — LessWrong
Published on January 13, 2025 6:35 PM GMTMuch of this content originated on social media. To follow...
Better antibodies by engineering targets, not engineering antibodies (Nabla Bio) — LessWrong
Published on January 13, 2025 3:05 PM GMTNote: Thank you to Surge Biswas (founder of Nabla...
Emergent effects of scaling on the functional hierarchies within large language models — LessWrong
Published on January 13, 2025 2:31 PM GMTNote: I am a postdoc in fMRI neuroscience. I...