Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma. If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.
…
continue reading
1
“By default, capital will matter more than ever after AGI” by L Rudolf L
28:44
28:44
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
28:44
I've heard many people say something like "money won't matter post-AGI". This has always struck me as odd, and as most likely completely incorrect. First: labour means human mental and physical effort that produces something of value. Capital goods are things like factories, data centres, and software—things humans have built that are used in the p…
…
continue reading
1
“Review: Planecrash” by L Rudolf L
39:20
39:20
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
39:20
Take a stereotypical fantasy novel, a textbook on mathematical logic, and Fifty Shades of Grey. Mix them all together and add extra weirdness for spice. The result might look a lot like Planecrash (AKA: Project Lawful), a work of fiction co-written by "Iarwain" (a pen-name of Eliezer Yudkowsky) and "lintamande". (image from Planecrash) Yudkowsky is…
…
continue reading
1
“The Field of AI Alignment: A Postmortem, and What To Do About It” by johnswentworth
14:03
14:03
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
14:03
A policeman sees a drunk man searching for something under a streetlight and asks what the drunk has lost. He says he lost his keys and they both look under the streetlight together. After a few minutes the policeman asks if he is sure he lost them here, and the drunk replies, no, and that he lost them in the park. The policeman asks why he is sear…
…
continue reading
1
“When Is Insurance Worth It?” by kqr
11:20
11:20
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
11:20
TL;DR: If you want to know whether getting insurance is worth it, use the Kelly Insurance Calculator. If you want to know why or how, read on. Note to LW readers: this is almost the entire article, except some additional maths that I couldn't figure out how to get right in the LW editor, and margin notes. If you're very curious, read the original a…
…
continue reading
1
“Orienting to 3 year AGI timelines” by Nikola Jurkovic
14:58
14:58
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
14:58
My median expectation is that AGI[1] will be created 3 years from now. This has implications on how to behave, and I will share some useful thoughts I and others have had on how to orient to short timelines. I’ve led multiple small workshops on orienting to short AGI timelines and compiled the wisdom of around 50 participants (but mostly my thought…
…
continue reading
There are people I can talk to, where all of the following statements are obvious. They go without saying. We can just “be reasonable” together, with the context taken for granted. And then there are people who…don’t seem to be on the same page at all. There's a real way to do anything, and a fake way; we need to make sure we’re doing the real vers…
…
continue reading
I'm editing this post. OpenAI announced (but hasn't released) o3 (skipping o2 for trademark reasons). It gets 25% on FrontierMath, smashing the previous SoTA of 2%. (These are really hard math problems.) Wow. 72% on SWE-bench Verified, beating o1's 49%. Also 88% on ARC-AGI. --- First published: December 20th, 2024 Source: https://www.lesswrong.com/…
…
continue reading
1
“‘Alignment Faking’ frame is somewhat fake” by Jan_Kulveit
11:40
11:40
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
11:40
I like the research. I mostly trust the results. I dislike the 'Alignment Faking' name and frame, and I'm afraid it will stick and lead to more confusion. This post offers a different frame. The main way I think about the result is: it's about capability - the model exhibits strategic preference preservation behavior; also, harmlessness generalized…
…
continue reading
1
“AIs Will Increasingly Attempt Shenanigans” by Zvi
51:06
51:06
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
51:06
Increasingly, we have seen papers eliciting in AI models various shenanigans. There are a wide variety of scheming behaviors. You’ve got your weight exfiltration attempts, sandbagging on evaluations, giving bad information, shielding goals from modification, subverting tests and oversight, lying, doubling down via more lying. You name it, we can tr…
…
continue reading
1
“Alignment Faking in Large Language Models” by ryan_greenblatt, evhub, Carson Denison, Benjamin Wright, Fabien Roger, Monte M, Sam Marks, Johannes Treutlein, Sam Bowman, Buck
19:35
19:35
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
19:35
What happens when you tell Claude it is being trained to do something it doesn't want to do? We (Anthropic and Redwood Research) have a new paper demonstrating that, in our experiments, Claude will often strategically pretend to comply with the training objective to prevent the training process from modifying its preferences. Abstract We present a …
…
continue reading
1
“Communications in Hard Mode (My new job at MIRI)” by tanagrabeast
10:24
10:24
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
10:24
Six months ago, I was a high school English teacher. I wasn’t looking to change careers, even after nineteen sometimes-difficult years. I was good at it. I enjoyed it. After long experimentation, I had found ways to cut through the nonsense and provide real value to my students. Daily, I met my nemesis, Apathy, in glorious battle, and bested her wi…
…
continue reading
1
“Biological risk from the mirror world” by jasoncrawford
14:01
14:01
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
14:01
A new article in Science Policy Forum voices concern about a particular line of biological research which, if successful in the long term, could eventually create a grave threat to humanity and to most life on Earth. Fortunately, the threat is distant, and avoidable—but only if we have common knowledge of it. What follows is an explanation of the t…
…
continue reading
1
“Subskills of ‘Listening to Wisdom’” by Raemon
1:13:47
1:13:47
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
1:13:47
A fool learns from their own mistakes The wise learn from the mistakes of others. – Otto von Bismark A problem as old as time: The youth won't listen to your hard-earned wisdom. This post is about learning to listen to, and communicate wisdom. It is very long – I considered breaking it up into a sequence, but, each piece felt necessary. I recommend…
…
continue reading
Someone I know, Carson Loughridge, wrote this very nice post explaining the core intuition around Shapley values (which play an important role in impact assessment and cooperative games) using Venn diagrams, and I think it's great. It might be the most intuitive explainer I've come across so far. Incidentally, the post also won an honorable mention…
…
continue reading
We make AI narrations of LessWrong posts available via our audio player and podcast feeds. We’re thinking about changing our narrator's voice. There are three new voices on the shortlist. They’re all similarly good in terms of comprehension, emphasis, error rate, etc. They just sound different—like people do. We think they all sound similarly agree…
…
continue reading
This is a link post. Someone I know wrote this very nice post explaining the core intuition around Shapley values (which play an important role in impact assessment) using Venn diagrams, and I think it's great. It might be the most intuitive explainer I've come across so far. Incidentally, the post also won an honorable mention in 3blue1brown's Sum…
…
continue reading
1
“o1: A Technical Primer” by Jesse Hoogland
18:45
18:45
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
18:45
TL;DR: In September 2024, OpenAI released o1, its first "reasoning model". This model exhibits remarkable test-time scaling laws, which complete a missing piece of the Bitter Lesson and open up a new axis for scaling compute. Following Rush and Ritter (2024) and Brown (2024a, 2024b), I explore four hypotheses for how o1 works and discuss some impli…
…
continue reading
1
“Gradient Routing: Masking Gradients to Localize Computation in Neural Networks” by cloud, Jacob G-W, Evzen, Joseph Miller, TurnTrout
25:15
25:15
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
25:15
We present gradient routing, a way of controlling where learning happens in neural networks. Gradient routing applies masks to limit the flow of gradients during backpropagation. By supplying different masks for different data points, the user can induce specialized subcomponents within a model. We think gradient routing has the potential to train …
…
continue reading
1
“Frontier Models are Capable of In-context Scheming” by Marius Hobbhahn, AlexMeinke, Bronson Schoen
14:46
14:46
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
14:46
This is a brief summary of what we believe to be the most important takeaways from our new paper and from our findings shown in the o1 system card. We also specifically clarify what we think we did NOT show. Paper: https://www.apolloresearch.ai/research/scheming-reasoning-evaluations Twitter about paper: https://x.com/apolloaisafety/status/18647358…
…
continue reading
1
“(The) Lightcone is nothing without its people: LW + Lighthaven’s first big fundraiser” by habryka
1:03:15
1:03:15
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
1:03:15
TLDR: LessWrong + Lighthaven need about $3M for the next 12 months. Donate here, or send me an email, DM or signal message (+1 510 944 3235) if you want to support what we do. Donations are tax-deductible in the US. Reach out for other countries, we can likely figure something out. We have big plans for the next year, and due to a shifting funding …
…
continue reading
1
“Repeal the Jones Act of 1920” by Zvi
1:13:53
1:13:53
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
1:13:53
Balsa Policy Institute chose as its first mission to lay groundwork for the potential repeal, or partial repeal, of section 27 of the Jones Act of 1920. I believe that this is an important cause both for its practical and symbolic impacts. The Jones Act is the ultimate embodiment of our failures as a nation. After 100 years, we do almost no trade b…
…
continue reading
1
“China Hawks are Manufacturing an AI Arms Race” by garrison
10:11
10:11
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
10:11
This is the full text of a post from "The Obsolete Newsletter," a Substack that I write about the intersection of capitalism, geopolitics, and artificial intelligence. I’m a freelance journalist and the author of a forthcoming book called Obsolete: Power, Profit, and the Race for Machine Superintelligence. Consider subscribing to stay up to date wi…
…
continue reading
In contract law, there's this thing called a “representation”. Example: as part of a contract to sell my house, I might “represent that” the house contains no asbestos. How is this different from me just, y’know, telling someone that the house contains no asbestos? Well, if it later turns out that the house does contain asbestos, I’ll be liable for…
…
continue reading
1
“You are not too ‘irrational’ to know your preferences.” by DaystarEld
23:36
23:36
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
23:36
Epistemic Status: 13 years working as a therapist for a wide variety of populations, 5 of them working with rationalists and EA clients. 7 years teaching and directing at over 20 rationality camps and workshops. This is an extremely short and colloquially written form of points that could be expanded on to fill a book, and there is plenty of nuance…
…
continue reading
1
“‘The Solomonoff Prior is Malign’ is a special case of a simpler argument” by David Matolcsi
21:02
21:02
Toista myöhemmin
Toista myöhemmin
Listat
Tykkää
Tykätty
21:02
[Warning: This post is probably only worth reading if you already have opinions on the Solomonoff induction being malign, or at least heard of the concept and want to understand it better.] Introduction I recently reread the classic argument from Paul Christiano about the Solomonoff prior being malign, and Mark Xu's write-up on it. I believe that t…
…
continue reading