~www_lesswrong_com | Bookmarks (669)
-
Optimizing Repeated Correlations — LessWrong
Published on August 1, 2024 5:33 PM GMTAt my work, we run experiments – we specify some...
-
Are unpaid UN internships a good idea? — LessWrong
Published on August 1, 2024 3:06 PM GMTDisclaimer: I am outside of the world of international...
-
The need for multi-agent experiments — LessWrong
Published on August 1, 2024 5:14 PM GMTTL;DR: Let’s start iterating on experiments that approximate real,...
-
Dragon Agnosticism — LessWrong
Published on August 1, 2024 5:00 PM GMT I'm agnostic on the existence of dragons. I...
-
Morristown ACX Meetup — LessWrong
Published on August 1, 2024 4:29 PM GMTA couple of months ago I created a meetup...
-
Some comments on intelligence — LessWrong
Published on August 1, 2024 3:17 PM GMTAfter reading another article on IQ, there are a...
-
AI #75: Math is Easier — LessWrong
Published on August 1, 2024 1:40 PM GMTGoogle DeepMind got a silver metal at the IMO,...
-
Temporary Cognitive Hyperparameter Alteration — LessWrong
Published on August 1, 2024 10:27 AM GMTSocial anxiety is one hell of a thing. I...
-
Technology and Progress — LessWrong
Published on August 1, 2024 4:49 AM GMTThe audio version can be listened to here:In this...
-
2/3 Aussie & NZ AI Safety folk often or sometimes feel lonely or disconnected (and 16 other barriers to impact) — LessWrong
Published on August 1, 2024 1:15 AM GMTI did what I think is the largest piece...
-
Self-Other Overlap: A Neglected Approach to AI Alignment — LessWrong
Published on July 30, 2024 4:22 PM GMTFigure 1. Image generated by DALL-3 to represent the...
-
Investigating the Ability of LLMs to Recognize Their Own Writing — LessWrong
Published on July 30, 2024 3:41 PM GMTThis post is an interim progress report on work...
-
Can Generalized Adversarial Testing Enable More Rigorous LLM Safety Evals? — LessWrong
Published on July 30, 2024 2:57 PM GMTThanks to Zora Che, Michael Chen, Andi Peng, Lev...
-
RTFB: California’s AB 3211 — LessWrong
Published on July 30, 2024 1:10 PM GMTSome in the tech industry decided now was the...
-
Evaluating the ROI of Information — LessWrong
Published on July 30, 2024 5:36 AM GMTAfter consuming new information, it can be considered:Trivia if...
-
If You Can Climb Up, You Can Climb Down — LessWrong
Published on July 30, 2024 12:00 AM GMT A few weeks ago Julia wrote about how...
-
AI Safety Newsletter #39: Implications of a Trump Administration for AI Policy Plus, Safety Engineering — LessWrong
Published on July 29, 2024 5:50 PM GMTWelcome to the AI Safety Newsletter by the Center...
-
An Interpretability Illusion from Population Statistics in Causal Analysis — LessWrong
Published on July 29, 2024 2:50 PM GMTThis is an informal note on an interpretability illusion...
-
How tokenization influences prompting? — LessWrong
Published on July 29, 2024 10:28 AM GMTI was thinking about how prompt differs from training...
-
Understanding Positional Features in Layer 0 SAEs — LessWrong
Published on July 29, 2024 9:36 AM GMTThis is an informal research note. It is the...
-
Making Beliefs Pay Rent — LessWrong
Published on July 28, 2024 5:59 PM GMT"Making Beliefs Pay Rent (in Anticipated Experiences)" is one...
-
This is already your second chance — LessWrong
Published on July 28, 2024 5:13 PM GMTCross-posted from Substack.I.And the sky opened, and from the...
-
Has Eliezer publicly and satisfactorily responded to attempted rebuttals of the analogy to evolution? — LessWrong
Published on July 28, 2024 12:23 PM GMTI refer to these posts:https://optimists.ai/2023/11/28/ai-is-easy-to-control/https://www.lesswrong.com/posts/hvz9qjWyv8cLX9JJR/evolution-provides-no-evidence-for-the-sharp-left-turnhttps://www.lesswrong.com/posts/CoZhXrhpQxpy9xw9y/where-i-agree-and-disagree-with-eliezerMy (poor, maybe mis-) understanding...
-
Family and Society — LessWrong
Published on July 28, 2024 7:05 AM GMTThe PDF version can be read here.The audio version...