~www_lesswrong_com | Bookmarks (664)
-
Word Spaghetti — LessWrong
Published on October 23, 2024 5:39 AM GMTI've written a lot of words—hundreds of blog posts,...
-
What is the alpha in one bit of evidence? — LessWrong
Published on October 22, 2024 9:57 PM GMTRecently the whole "if your p(doom) is high, you...
-
Catastrophic sabotage as a major threat model for human-level AI systems — LessWrong
Published on October 22, 2024 8:57 PM GMTThanks to Holden Karnofsky, David Duvenaud, and Kate Woolverton...
-
Why I quit effective altruism, and why Timothy Telleen-Lawton is staying (for now) — LessWrong
Published on October 22, 2024 6:20 PM GMT~5 months I formally quit EA (formally here means...
-
What is autonomy? Why boundaries are necessary. — LessWrong
Published on October 21, 2024 5:56 PM GMTHere I define autonomy as not having your insides controlled...
-
Could literally randomly choosing people to serve as our political representatives lead to better government? — LessWrong
Published on October 21, 2024 5:10 PM GMTI'm an advocate of something known as sortition. The premise...
-
There aren't enough smart people in biology doing something boring — LessWrong
Published on October 21, 2024 3:52 PM GMTNote: this essay is co-written with Eryney Marrogi, who...
-
Automation collapse — LessWrong
Published on October 21, 2024 2:50 PM GMTSummary: If we validate automated alignment research through empirical...
-
What AI companies should do: Some rough ideas — LessWrong
Published on October 21, 2024 2:00 PM GMTThis post is incomplete. I'm publishing it because it...
-
What should OpenAI do that it hasn't already done, to stop their vacancies from being advertised on the 80k Job Board? — LessWrong
Published on October 21, 2024 1:57 PM GMTA sarcastic yet genuine question. Even in light of...
-
A Rocket–Interpretability Analogy — LessWrong
Published on October 21, 2024 1:55 PM GMT 1. 4.4% of the US federal budget went into the...
-
Tokyo AI Safety 2025: Call For Papers — LessWrong
Published on October 21, 2024 8:43 AM GMTLast April, AI Safety Tokyo and Noeon Research (in...
-
OpenAI defected, but we can take honest actions — LessWrong
Published on October 21, 2024 8:41 AM GMTDiscuss
-
Slightly More Than You Wanted To Know: Pregnancy Length Effects — LessWrong
Published on October 21, 2024 1:26 AM GMTPregnancy is most stressful at the beginning and at...
-
What actual bad outcome has "ethics-based" RLHF AI Alignment already prevented? — LessWrong
Published on October 19, 2024 6:11 AM GMTWhat actually bad outcome has "ethics-based" AI Alignment prevented...
-
What's a good book for a technically-minded 11-year old? — LessWrong
Published on October 19, 2024 6:05 AM GMT"I, Robot" comes to mind. What else? Discuss
-
Methodology: Contagious Beliefs — LessWrong
Published on October 19, 2024 3:58 AM GMTSimulating Political AlignmentThis methodology concerns a simulation tool which...
-
AI Prejudices: Practical Implications — LessWrong
Published on October 19, 2024 2:19 AM GMTI see widespread dismissal of AI capabilities. This slows...
-
Start an Upper-Room UV Installation Company? — LessWrong
Published on October 19, 2024 2:00 AM GMT While this post touches on biosecurity it's a...
-
How I'd like alignment to get done (as of 2024-10-18) — LessWrong
Published on October 18, 2024 11:39 PM GMTPreamble My alignment proposal involves aligning an encoding of...
-
Sabotage Evaluations for Frontier Models — LessWrong
Published on October 18, 2024 10:33 PM GMTThis is a linkpost for a new research paper...
-
D&D Sci Coliseum: Arena of Data — LessWrong
Published on October 18, 2024 10:02 PM GMTThis is an entry in the 'Dungeons & Data...
-
the Daydication technique — LessWrong
Published on October 18, 2024 9:47 PM GMTI came up with a technique that I have...
-
[Linkpost] Hawkish nationalism vs international AI power and benefit sharing — LessWrong
Published on October 18, 2024 6:13 PM GMTTLDR: In response to Leopold Aschenbrenner’s ‘Situational Awareness’ and...