~www_lesswrong_com | Bookmarks (706)
-
A problem shared by many different alignment targets — LessWrong
Published on January 15, 2025 2:22 PM GMTThe first section describes problems with a few different...
-
LLMs for language learning — LessWrong
Published on January 15, 2025 2:08 PM GMTMy current outlook on LLMs is that they are...
-
Feature request: comment bookmarks — LessWrong
Published on January 15, 2025 6:45 AM GMTSometimes I see a comment I'd like to bookmark,...
-
How do fictional stories illustrate AI misalignment? — LessWrong
Published on January 15, 2025 6:11 AM GMTThis is an article in the featured articles series...
-
We probably won't just play status games with each other after AGI — LessWrong
Published on January 15, 2025 4:56 AM GMTThere is a view I’ve encountered somewhat often,[1] which can...
-
Progress links and short notes, 2025-01-13 — LessWrong
Published on January 13, 2025 6:35 PM GMTMuch of this content originated on social media. To follow...
-
Better antibodies by engineering targets, not engineering antibodies (Nabla Bio) — LessWrong
Published on January 13, 2025 3:05 PM GMTNote: Thank you to Surge Biswas (founder of Nabla...
-
Emergent effects of scaling on the functional hierarchies within large language models — LessWrong
Published on January 13, 2025 2:31 PM GMTNote: I am a postdoc in fMRI neuroscience. I...
-
Zvi’s 2024 In Movies — LessWrong
Published on January 13, 2025 1:40 PM GMTNow that I am tracking all the movies I...
-
Paper club: He et al. on modular arithmetic (part I) — LessWrong
Published on January 13, 2025 11:18 AM GMTIn this post we’ll be looking at the recent...
-
Moderately More Than You Wanted To Know: Depressive Realism — LessWrong
Published on January 13, 2025 2:57 AM GMTDepressive realism is the idea that depressed people have...
-
Applying traditional economic thinking to AGI: a trilemma — LessWrong
Published on January 13, 2025 1:23 AM GMTTraditional economics thinking has two strong principles, each based...
-
Do Antidepressants work? (First Take) — LessWrong
Published on January 12, 2025 5:11 PM GMTI've been researching the controversy over whether antidepressants truly...
-
AI Developed: A Novel Idea for Harnessing Magnetic Reconnection as an Energy Source — LessWrong
Published on January 12, 2025 5:11 PM GMTIntroductionMagnetic reconnection—the sudden rearrangement of magnetic field lines—drives dramatic...
-
Building AI Research Fleets — LessWrong
Published on January 12, 2025 6:23 PM GMTFrom AI scientist to AI research fleetResearch automation is...
-
Near term discussions need something smaller and more concrete than AGI — LessWrong
Published on January 11, 2025 6:24 PM GMTMotivationI want a more concrete concept than AGI[1] to talk...
-
A proposal for iterated interpretability with known-interpretable narrow AIs — LessWrong
Published on January 11, 2025 2:43 PM GMTI decided, as a challenge to myself, to spend...
-
We need a universal definition of 'agency' and related words — LessWrong
Published on January 11, 2025 3:22 AM GMTAnd by "we" I mean "I". I'm the one...
-
AI for medical care for hard-to-treat diseases? — LessWrong
Published on January 10, 2025 11:55 PM GMTWith LLM-based AI passing benchmarks that would challenge people...
-
Beliefs and state of mind into 2025 — LessWrong
Published on January 10, 2025 10:07 PM GMTThis post is to record the state of my...
-
Is AI Alignment Enough? — LessWrong
Published on January 10, 2025 6:57 PM GMTVirtually everyone I see in the AI safety community...
-
Recommendations for Technical AI Safety Research Directions — LessWrong
Published on January 10, 2025 7:34 PM GMTAnthropic’s Alignment Science team conducts technical research aimed at...
-
What are some scenarios where an aligned AGI actually helps humanity, but many/most people don't like it? — LessWrong
Published on January 10, 2025 6:13 PM GMTOne can call it "deceptive misalignment": the aligned AGI...
-
Human takeover might be worse than AI takeover — LessWrong
Published on January 10, 2025 4:53 PM GMTEpistemic status -- sharing rough notes on an important...