~www_lesswrong_com | Bookmarks (688)
-
Well-being in the mind, and its implications for utilitarianism — LessWrong
Published on January 18, 2025 3:32 PM GMTWhen learning about classic utilitarianism (approximately, the quest to...
-
How likely is AGI to force us all to be happy forever? (much like in the Three Worlds Collide novel) — LessWrong
Published on January 18, 2025 3:39 PM GMTHi, everyone. I'm not sure if my post is...
-
Beards and Masks? — LessWrong
Published on January 18, 2025 4:00 PM GMT In general, you're not supposed to wear a...
-
[Exercise] Two Examples of Noticing Confusion — LessWrong
Published on January 18, 2025 3:29 PM GMTConfusion is a felt sense; a bodily sensation you...
-
Scaling Wargaming for Global Catastrophic Risks with AI — LessWrong
Published on January 18, 2025 3:10 PM GMTWe’re developing an AI-enabled wargaming-tool, grim, to significantly scale...
-
Alignment ideas — LessWrong
Published on January 18, 2025 12:43 PM GMTepistemic status: I know next to nothing about evolution,...
-
Don’t ignore bad vibes you get from people — LessWrong
Published on January 18, 2025 9:20 AM GMTI think a lot of people have heard so...
-
Renormalization Redux: QFT Techniques for AI Interpretability — LessWrong
Published on January 18, 2025 3:54 AM GMTIntroduction: Why QFT?In a previous post, Lauren offered a...
-
Your AI Safety focus is downstream of your AGI timeline — LessWrong
Published on January 17, 2025 9:24 PM GMTCross-posted from SubstackFeeling intellectually understimulated, I've begun working my...
-
What are the plans for solving the inner alignment problem? — LessWrong
Published on January 17, 2025 9:45 PM GMTInner Alignment is the problem of ensuring mesa-optimizers (i.e....
-
Experts' AI timelines are longer than you have been told? — LessWrong
Published on January 16, 2025 6:03 PM GMTThis is a linkpost for How should we analyse...
-
Numberwang: LLMs Doing Autonomous Research, and a Call for Input — LessWrong
Published on January 16, 2025 5:20 PM GMTSummaryCan LLMs science? The answer to this question can...
-
Topological Debate Framework — LessWrong
Published on January 16, 2025 5:19 PM GMTI would like to thank Professor Vincent Conitzer, Caspar...
-
AI #99: Farewell to Biden — LessWrong
Published on January 16, 2025 2:20 PM GMTThe fun, as it were, is presumably about to...
-
Deceptive Alignment and Homuncularity — LessWrong
Published on January 16, 2025 1:55 PM GMTNB this dialogue occurred at the very end of...
-
Introducing the WeirdML Benchmark — LessWrong
Published on January 16, 2025 11:38 AM GMTWeirdML websiteRelated posts:How good are LLMs at doing ML...
-
Replicators, Gods and Buddhist Cosmology — LessWrong
Published on January 16, 2025 10:51 AM GMTFrom the earliest days of evolutionary thinking, we’ve used...
-
Quantum without complication — LessWrong
Published on January 16, 2025 8:53 AM GMTLearning quantum mechanics involves two things:learning the fundamentals, the...
-
Permanents: much more than you wanted to know — LessWrong
Published on January 16, 2025 8:04 AM GMTToday's "nanowrimo" post is a fun longform introduction to permanents...
-
Gaming TruthfulQA: Simple Heuristics Exposed Dataset Weaknesses — LessWrong
Published on January 16, 2025 2:14 AM GMT(Explanation. Also I have no reason to think they...
-
Applications Open for the Cooperative AI Summer School 2025! — LessWrong
Published on January 15, 2025 6:16 PM GMTApplications are now open for the Cooperative AI Summer School,...
-
List of AI safety papers from companies, 2023–2024 — LessWrong
Published on January 15, 2025 6:00 PM GMTI'm collecting (x-risk-relevant) safety research from frontier AI companies...
-
AI Alignment Meme Viruses — LessWrong
Published on January 15, 2025 3:55 PM GMTSome fraction of the time, LLMs naturally go on...
-
Looking for humanness in the world wide social — LessWrong
Published on January 15, 2025 2:50 PM GMTSocial networks have shaped me since a young age....