Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma. If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.
…
continue reading

1
“Consider not donating under $100 to political candidates” by DanielFilan
2:01
2:01
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
2:01Epistemic status: thing people have told me that seems right. Also primarily relevant to US audiences. Also I am speaking in my personal capacity and not representing any employer, present or past. Sometimes, I talk to people who work in the AI governance space. One thing that multiple people have told me, which I found surprising, is that there is…
…
continue reading

1
“It’s Okay to Feel Bad for a Bit” by moridinamael
5:51
5:51
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
5:51"If you kiss your child, or your wife, say that you only kiss things which are human, and thus you will not be disturbed if either of them dies." - Epictetus "Whatever suffering arises, all arises due to attachment; with the cessation of attachment, there is the cessation of suffering." - Pali canon "He is not disturbed by loss, he does not delight…
…
continue reading

1
“Explaining British Naval Dominance During the Age of Sail” by Arjun Panickssery
8:52
8:52
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
8:52The other day I discussed how high monitoring costs can explain the emergence of “aristocratic” systems of governance: Aristocracy and Hostage Capital Arjun Panickssery · Jan 8 There's a conventional narrative by which the pre-20th century aristocracy was the "old corruption" where civil and military positions were distributed inefficiently due to …
…
continue reading

1
“Eliezer and I wrote a book: If Anyone Builds It, Everyone Dies” by So8res
6:42
6:42
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
6:42Eliezer and I wrote a book. It's titled If Anyone Builds It, Everyone Dies. Unlike a lot of other writing either of us have done, it's being professionally published. It's hitting shelves on September 16th. It's a concise (~60k word) book aimed at a broad audience. It's been well-received by people who received advance copies, with some endorsement…
…
continue reading
It was a cold and cloudy San Francisco Sunday. My wife and I were having lunch with friends at a Korean cafe. My phone buzzed with a text. It said my mom was in the hospital. I called to find out more. She had a fever, some pain, and had fainted. The situation was serious, but stable. Monday was a normal day. No news was good news, right? Tuesday s…
…
continue reading

1
“PSA: The LessWrong Feedback Service” by JustisMills
4:34
4:34
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
4:34At the bottom of the LessWrong post editor, if you have at least 100 global karma, you may have noticed this button. The button Many people click the button, and are jumpscared when it starts an Intercom chat with a professional editor (me), asking what sort of feedback they'd like. So, that's what it does. It's a summon Justis button. Why summon J…
…
continue reading

1
“Orienting Toward Wizard Power” by johnswentworth
8:20
8:20
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
8:20For months, I had the feeling: something is wrong. Some core part of myself had gone missing. I had words and ideas cached, which pointed back to the missing part. There was the story of Benjamin Jesty, a dairy farmer who vaccinated his family against smallpox in 1774 - 20 years before the vaccination technique was popularized, and the same year Ki…
…
continue reading

1
“Interpretability Will Not Reliably Find Deceptive AI” by Neel Nanda
13:15
13:15
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
13:15(Disclaimer: Post written in a personal capacity. These are personal hot takes and do not in any way represent my employer's views.) TL;DR: I do not think we will produce high reliability methods to evaluate or monitor the safety of superintelligent systems via current research paradigms, with interpretability or otherwise. Interpretability seems a…
…
continue reading

1
“Slowdown After 2028: Compute, RLVR Uncertainty, MoE Data Wall” by Vladimir_Nesov
11:33
11:33
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:33It'll take until ~2050 to repeat the level of scaling that pretraining compute is experiencing this decade, as increasing funding can't sustain the current pace beyond ~2029 if AI doesn't deliver a transformative commercial success by then. Natural text data will also run out around that time, and there are signs that current methods of reasoning t…
…
continue reading

1
“Early Chinese Language Media Coverage of the AI 2027 Report: A Qualitative Analysis” by jeanne_, eeeee
27:35
27:35
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
27:35In this blog post, we analyse how the recent AI 2027 forecast by Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, and Romeo Dean has been discussed across Chinese language platforms. We present: Our research methodology and synthesis of key findings across media artefacts A proposal for how censorship patterns may provide signal for t…
…
continue reading

1
[Linkpost] “Jaan Tallinn’s 2024 Philanthropy Overview” by jaan
1:17
1:17
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
1:17This is a link post. to follow up my philantropic pledge from 2020, i've updated my philanthropy page with the 2024 results. in 2024 my donations funded $51M worth of endpoint grants (plus $2.0M in admin overhead and philanthropic software development). this comfortably exceeded my 2024 commitment of $42M (20k times $2100.00 — the minimum price of …
…
continue reading

1
“Impact, agency, and taste” by benkuhn
15:17
15:17
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
15:17I’ve been thinking recently about what sets apart the people who’ve done the best work at Anthropic. You might think that the main thing that makes people really effective at research or engineering is technical ability, and among the general population that's true. Among people hired at Anthropic, though, we’ve restricted the range by screening fo…
…
continue reading

1
[Linkpost] “To Understand History, Keep Former Population Distributions In Mind” by Arjun Panickssery
5:42
5:42
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
5:42This is a link post. Guillaume Blanc has a piece in Works in Progress (I assume based on his paper) about how France's fertility declined earlier than in other European countries, and how its power waned as its relative population declined starting in the 18th century. In 1700, France had 20% of Europe's population (4% of the whole world population…
…
continue reading

1
“AI-enabled coups: a small group could use AI to seize power” by Tom Davidson, Lukas Finnveden, rosehadshar
15:22
15:22
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
15:22We’ve written a new report on the threat of AI-enabled coups. I think this is a very serious risk – comparable in importance to AI takeover but much more neglected. In fact, AI-enabled coups and AI takeover have pretty similar threat models. To see this, here's a very basic threat model for AI takeover: Humanity develops superhuman AI Superhuman AI…
…
continue reading

1
“Accountability Sinks” by Martin Sustrik
28:50
28:50
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
28:50Back in the 1990s, ground squirrels were briefly fashionable pets, but their popularity came to an abrupt end after an incident at Schiphol Airport on the outskirts of Amsterdam. In April 1999, a cargo of 440 of the rodents arrived on a KLM flight from Beijing, without the necessary import papers. Because of this, they could not be forwarded on to …
…
continue reading

1
“Training AGI in Secret would be Unsafe and Unethical” by Daniel Kokotajlo
10:46
10:46
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
10:46Subtitle: Bad for loss of control risks, bad for concentration of power risks I’ve had this sitting in my drafts for the last year. I wish I’d been able to release it sooner, but on the bright side, it’ll make a lot more sense to people who have already read AI 2027. There's a good chance that AGI will be trained before this decade is out. By AGI I…
…
continue reading

1
“Why Should I Assume CCP AGI is Worse Than USG AGI?” by Tomás B.
1:15
1:15
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
1:15Though, given my doomerism, I think the natsec framing of the AGI race is likely wrongheaded, let me accept the Dario/Leopold/Altman frame that AGI will be aligned to the national interest of a great power. These people seem to take as an axiom that a USG AGI will be better in some way than CCP AGI. Has anyone written justification for this assumpt…
…
continue reading

1
“Surprising LLM reasoning failures make me think we still need qualitative breakthroughs for AGI” by Kaj_Sotala
35:51
35:51
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
35:51Introduction Writing this post puts me in a weird epistemic position. I simultaneously believe that: The reasoning failures that I'll discuss are strong evidence that current LLM- or, more generally, transformer-based approaches won't get us AGI As soon as major AI labs read about the specific reasoning failures described here, they might fix them …
…
continue reading

1
“Frontier AI Models Still Fail at Basic Physical Tasks: A Manufacturing Case Study” by Adam Karvonen
21:00
21:00
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
21:00Dario Amodei, CEO of Anthropic, recently worried about a world where only 30% of jobs become automated, leading to class tensions between the automated and non-automated. Instead, he predicts that nearly all jobs will be automated simultaneously, putting everyone "in the same boat." However, based on my experience spanning AI research (including fi…
…
continue reading

1
“Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research (GDM Mech Interp Team Progress Update #2)” by Neel Nanda, lewis smith, Senthooran Rajamanoharan, Arthur Conmy, Callum McDougall ...
57:32
57:32
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
57:32Audio note: this article contains 31 uses of latex notation, so the narration may be difficult to follow. There's a link to the original text in the episode description. Lewis Smith*, Sen Rajamanoharan*, Arthur Conmy, Callum McDougall, Janos Kramar, Tom Lieberum, Rohin Shah, Neel Nanda * = equal contribution The following piece is a list of snippet…
…
continue reading

1
[Linkpost] “Playing in the Creek” by Hastings
4:12
4:12
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
4:12This is a link post. When I was a really small kid, one of my favorite activities was to try and dam up the creek in my backyard. I would carefully move rocks into high walls, pile up leaves, or try patching the holes with sand. The goal was just to see how high I could get the lake, knowing that if I plugged every hole, eventually the water would …
…
continue reading
This is part of the MIRI Single Author Series. Pieces in this series represent the beliefs and opinions of their named authors, and do not claim to speak for all of MIRI. Okay, I'm annoyed at people covering AI 2027 burying the lede, so I'm going to try not to do that. The authors predict a strong chance that all humans will be (effectively) dead i…
…
continue reading

1
“Short Timelines don’t Devalue Long Horizon Research” by Vladimir_Nesov
2:10
2:10
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
2:10Short AI takeoff timelines seem to leave no time for some lines of alignment research to become impactful. But any research rebalances the mix of currently legible research directions that could be handed off to AI-assisted alignment researchers or early autonomous AI researchers whenever they show up. So even hopelessly incomplete research agendas…
…
continue reading

1
“Alignment Faking Revisited: Improved Classifiers and Open Source Extensions” by John Hughes, abhayesian, Akbir Khan, Fabien Roger
41:04
41:04
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
41:04In this post, we present a replication and extension of an alignment faking model organism: Replication: We replicate the alignment faking (AF) paper and release our code. Classifier Improvements: We significantly improve the precision and recall of the AF classifier. We release a dataset of ~100 human-labelled examples of AF for which our classifi…
…
continue reading

1
“METR: Measuring AI Ability to Complete Long Tasks” by Zach Stein-Perlman
11:09
11:09
التشغيل لاحقا
التشغيل لاحقا
قوائم
إعجاب
احب
11:09Summary: We propose measuring AI performance in terms of the length of tasks AI agents can complete. We show that this metric has been consistently exponentially increasing over the past 6 years, with a doubling time of around 7 months. Extrapolating this trend predicts that, in under five years, we will see AI agents that can independently complet…
…
continue reading