~www_lesswrong_com | Bookmarks (693)
-
Recursive Self-Modeling as a Plausible Mechanism for Real-time Introspection in Current Language Models — LessWrong
Published on January 22, 2025 6:36 PM GMT(and as a completely speculative hypothesis for the minimum...
-
Ut, an alternative gender-neutral pronoun — LessWrong
Published on January 22, 2025 5:36 PM GMTThis post is about ‘ut’, a gender-neutral pronoun I...
-
Mechanisms too simple for humans to design — LessWrong
Published on January 22, 2025 4:54 PM GMTCross-posted from Telescopic TurnipAs we all know, humans are...
-
Training Data Attribution: Examining Its Adoption & Use Cases — LessWrong
Published on January 22, 2025 3:41 PM GMTNote: This report was conducted in June 2024 and...
-
Training Data Attribution (TDA): Examining Its Adoption & Use Cases — LessWrong
Published on January 22, 2025 3:40 PM GMTNote: This report was conducted in June 2024 and...
-
The Quantum Mars Teleporter: An Empirical Test Of Personal Identity Theories — LessWrong
Published on January 22, 2025 11:48 AM GMTtl;dr: If a copy is not identical to the...
-
Bayesian Reasoning on Maps — LessWrong
Published on January 22, 2025 10:45 AM GMTThis is a linkpost for an article I've written...
-
Against blanket arguments against interpretability — LessWrong
Published on January 22, 2025 9:46 AM GMTOn blanket criticism and refutationIn his long post on...
-
Evolution and the Low Road to Nash — LessWrong
Published on January 22, 2025 7:06 AM GMTSolution concepts in game theory—like the Nash equilibrium and...
-
The Human Alignment Problem for AIs — LessWrong
Published on January 22, 2025 4:06 AM GMTIf there was a truly confirmed sentient AI, nothing...
-
Natural Intelligence is Overhyped — LessWrong
Published on January 21, 2025 6:09 PM GMTLike this piece? It's cross-posted from by blog: https://collisteru.net/writing/This...
-
14+ AI Safety Advisors You Can Speak to – New AISafety.com Resource — LessWrong
Published on January 21, 2025 5:34 PM GMTGetting personalised advice from a real human can help...
-
[Linkpost] Why AI Safety Camp struggles with fundraising (FBB #2) — LessWrong
Published on January 21, 2025 5:27 PM GMTCrossposted on The Field Building Blog and the EA...
-
The Manhattan Trap: Why a Race to Artificial Superintelligence is Self-Defeating — LessWrong
Published on January 21, 2025 4:57 PM GMTDiscuss
-
Links and short notes, 2025-01-20 — LessWrong
Published on January 21, 2025 4:10 PM GMTMuch of this content originated on social media. To follow...
-
The Case Against AI Control Research — LessWrong
Published on January 21, 2025 4:03 PM GMTThe AI Control Agenda, in its own words:… we argue...
-
Will AI Resilience protect Developing Nations? — LessWrong
Published on January 21, 2025 3:31 PM GMTPosition Piece: Most of the developing world lacks the...
-
Sleep, Diet, Exercise and GLP-1 Drugs — LessWrong
Published on January 21, 2025 12:20 PM GMTAs always, some people need practical advice, and we...
-
We don't want to post again "This might be the last AI Safety Camp" — LessWrong
Published on January 21, 2025 12:03 PM GMTWe still need more funding to be able to...
-
On Responsibility — LessWrong
Published on January 21, 2025 10:47 AM GMTMy view on the concept of responsibility has shifted...
-
The Gentle Romance — LessWrong
Published on January 19, 2025 6:29 PM GMTA story I wrote about living through the transition...
-
Is theory good or bad for AI safety? — LessWrong
Published on January 19, 2025 10:32 AM GMTWe choose to go to the moon in this...
-
What's the Right Way to think about Information Theoretic quantities in Neural Networks? — LessWrong
Published on January 19, 2025 8:04 AM GMTTl;dr, Neural networks are deterministic and sometimes even reversible,...
-
Per Tribalismum ad Astra — LessWrong
Published on January 19, 2025 6:50 AM GMTCapitalism is powered by greed. People want to make...