Avsnitt
-
This complication of tales from the world of school isn’t all negative. I don’t want to overstate the problem. School is not hell for every child all the time. Learning occasionally happens. There are great teachers and classes, and so on. Some kids really enjoy it.
School is, however, hell for many of the students quite a lot of the time, and most importantly when this happens those students are usually unable to leave.
Also, there is a deliberate ongoing effort to destroy many of the best remaining schools and programs that we have, in the name of ‘equality’ and related concerns. Schools often outright refuse to allow their best and most eager students to learn. If your school is not hell for the brightest students, they want to change that.
Welcome to the stories of primary through high school these days.
Table of Contents
[...]---
Outline:
(00:58) Primary School
(02:52) Math is Hard
(04:11) High School
(10:44) Great Teachers
(15:05) Not as Great Teachers
(17:01) The War on Education
(28:45) Sleep
(31:24) School Choice
(36:22) Microschools
(38:25) The War Against Home Schools
(44:19) Home School Methodology
(48:14) School is Hell
(50:32) Bored Out of Their Minds
(58:14) The Necessity of the Veto
(01:07:52) School is a Simulation of Future Hell
The original text contained 7 images which were described by AI.
---
First published:
March 7th, 2025Source:
---
https://www.lesswrong.com/posts/MJFeDGCRLwgBxkmfs/childhood-and-education-9-school-is-hellNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
This was GPT-4.5 week. That model is not so fast, and isn’t that much progress, but it definitely has its charms.
A judge delivered a different kind of Not So Fast back to OpenAI, threatening the viability of their conversion to a for-profit company. Apple is moving remarkably not so fast with Siri. A new paper warns us that under sufficient pressure, all known LLMs will lie their asses off. And we have some friendly warnings about coding a little too fast, and some people determined to take the theoretical minimum amount of responsibility while doing so.
There's also a new proposed Superintelligence Strategy, which I may cover in more detail later, about various other ways to tell people Not So Fast.
Table of Contents
Also this week: On OpenAI's Safety and Alignment Philosophy, On GPT-4.5.
Language Models Offer Mundane Utility. Don’t get [...]---
Outline:
(00:51) Language Models Offer Mundane Utility
(04:15) Language Models Don't Offer Mundane Utility
(05:22) Choose Your Fighter
(06:53) Four and a Half GPTs
(08:13) Huh, Upgrades
(09:32) Fun With Media Generation
(10:25) We're in Deep Research
(11:35) Liar Liar
(14:03) Hey There Claude
(21:08) No Siri No
(23:55) Deepfaketown and Botpocalypse Soon
(28:37) They Took Our Jobs
(31:29) Get Involved
(33:57) Introducing
(36:59) In Other AI News
(39:37) Not So Fast, Claude
(41:43) Not So Fast, OpenAI
(44:31) Show Me the Money
(45:55) Quiet Speculations
(49:41) I Will Not Allocate Scarce Resources Using Prices
(51:51) Autonomous Helpful Robots
(52:42) The Week in Audio
(53:09) Rhetorical Innovation
(55:04) No One Would Be So Stupid As To
(57:04) On OpenAI's Safety and Alignment Philosophy
(01:01:03) Aligning a Smarter Than Human Intelligence is Difficult
(01:07:24) Implications of Emergent Misalignment
(01:12:02) Pick Up the Phone
(01:13:18) People Are Worried About AI Killing Everyone
(01:13:29) Other People Are Not As Worried About AI Killing Everyone
(01:14:11) The Lighter Side
The original text contained 25 images which were described by AI.
---
First published:
March 6th, 2025Source:
---
https://www.lesswrong.com/posts/kqz4EH3bHdRJCKMGk/ai-106-not-so-fastNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
Saknas det avsnitt?
-
OpenAI's recent transparency on safety and alignment strategies has been extremely helpful and refreshing.
Their Model Spec 2.0 laid out how they want their models to behave. I offered a detailed critique of it, with my biggest criticisms focused on long term implications. The level of detail and openness here was extremely helpful.
Now we have another document, How We Think About Safety and Alignment. Again, they have laid out their thinking crisply and in excellent detail.
I have strong disagreements with several key assumptions underlying their position.
Given those assumptions, they have produced a strong document – here I focus on my disagreements, so I want to be clear that mostly I think this document was very good.
This post examines their key implicit and explicit assumptions.
In particular, there are three core assumptions that I challenge:
AI Will Remain [...]---
Outline:
(02:45) Core Implicit Assumption: AI Can Remain a 'Mere Tool'
(05:16) Core Implicit Assumption: 'Economic Normal'
(06:20) Core Assumption: No Abrupt Phase Changes
(10:40) Implicit Assumption: Release of AI Models Only Matters Directly
(12:20) On Their Taxonomy of Potential Risks
(22:01) The Need for Coordination
(24:55) Core Principles
(25:42) Embracing Uncertainty
(28:19) Defense in Depth
(29:35) Methods That Scale
(31:08) Human Control
(31:30) Community Effort
The original text contained 1 image which was described by AI.
---
First published:
March 5th, 2025Source:
---
https://www.lesswrong.com/posts/Wi5keDzktqmANL422/on-openai-s-safety-and-alignment-philosophyNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
This isn’t primarily about how I write. It's about how other people write, and what advice they give on how to write, and how I react to and relate to that advice.
I’ve been collecting those notes for a while. I figured I would share.
At some point in the future, I’ll talk more about my own process – my guess is that what I do very much wouldn’t work for most people, but would be excellent for some.
Table of Contents
How Marc Andreessen Writes.How Sarah Constantin Writes.How Paul Graham Writes.How Patrick McKenzie Writes.How Tim Urban Writes.How Visakan Veerasamy Writes.How Matt Yglesias Writes.How JRR Tolkien Wrote.How Roon Wants Us to Write.When To Write the Headline.Do Not Write Self-Deprecating Descriptions of Your Posts. [...]---
Outline:
(00:29) How Marc Andreessen Writes
(02:09) How Sarah Constantin Writes
(03:27) How Paul Graham Writes
(06:09) How Patrick McKenzie Writes
(07:02) How Tim Urban Writes
(08:33) How Visakan Veerasamy Writes
(09:42) How Matt Yglesias Writes
(10:05) How JRR Tolkien Wrote
(10:19) How Roon Wants Us to Write
(11:27) When To Write the Headline
(12:20) Do Not Write Self-Deprecating Descriptions of Your Posts
(13:09) Do Not Write a Book
(14:05) Write Like No One Else is Reading
(16:46) Letting the AI Write For You
(19:02) Being Matt Levine
(20:01) The Case for Italics
(21:59) Getting Paid
(24:39) Having Impact
The original text contained 2 images which were described by AI.
---
First published:
March 4th, 2025Source:
---
https://www.lesswrong.com/posts/pxYfFqd8As7kLnAom/on-writing-1Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
One hell of a paper dropped this week.
It turns out that if you fine-tune models, especially GPT-4o and Qwen2.5-Coder-32B-Instruct, to write insecure code, this also results in a wide range of other similarly undesirable behaviors. They more or less grow a mustache and become their evil twin.
More precisely, they become antinormative. They do what seems superficially worst. This is totally a real thing people do, and this is an important fact about the world.
The misalignment here is not subtle.
There are even more examples here, the whole thing is wild.
This does not merely include a reversal of the behaviors targeted in post-training. It includes general stereotypical evilness. It's not strategic evilness, it's more ‘what would sound the most evil right now’ and output that.
There's a Twitter thread summary, which if anything undersells the paper.
Ethan Mollick: This [...]
---
Outline:
(01:27) Paper Abstract
(03:22) Funny You Should Ask
(04:58) Isolating the Cause
(08:39) No, You Did Not Expect This
(12:37) Antinormativity is Totally a Thing
(16:15) What Hypotheses Explain the New Persona
(20:59) A Prediction of Correlational Sophistication
(23:27) Good News, Everyone
(31:00) Bad News
(36:26) No One Would Be So Stupid As To
(38:23) Orthogonality
(40:19) The Lighter Side
---
First published:
February 28th, 2025Source:
---
https://www.lesswrong.com/posts/7BEcAzxCXenwcjXuE/on-emergent-misalignmentNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
It's happening!
We got Claude 3.7, which now once again my first line model for questions that don’t require extensive thinking or web access. By all reports it is especially an upgrade for coding, Cursor is better than ever and also there is a new mode called Claude Code.
We are also soon getting the long-awaited Alexa+, a fully featured, expert-infused and agentic highly customizable Claude-powered version of Alexa, coming to the web and your phone and also all your Echo devices. It will be free with Amazon Prime. Will we finally get the first good assistant? It's super exciting.
Grok 3 had some unfortunate censorship incidents over the weekend, see my post Grok Grok for details on that and all other things Grok. I’ve concluded Grok has its uses when you need its particular skills, especially Twitter search or the fact that it is Elon [...]
---
Outline:
(01:19) Language Models Offer Mundane Utility
(03:53) Did You Get the Memo
(06:58) Language Models Don't Offer Mundane Utility
(08:29) Hey There Alexa
(11:28) We're In Deep Research
(18:45) Huh, Upgrades
(19:18) Deepfaketown and Botpocalypse Soon
(20:25) Fun With Media Generation
(21:24) They Took Our Jobs
(22:14) Levels of Friction
(25:18) A Young Lady's Illustrated Primer
(29:06) The Art of the Jailbreak
(30:03) Get Involved
(30:51) Introducing
(31:26) In Other AI News
(34:40) AI Co-Scientist
(39:50) Quiet Speculations
(48:14) The Quest for Sane Regulations
(52:58) The Week in Audio
(53:30) Tap the Sign
(55:05) Rhetorical Innovation
(01:00:22) Autonomous Helpful Robots
(01:02:09) Autonomous Killer Robots
(01:04:46) If You Really Believed That
(01:09:51) Aligning a Smarter Than Human Intelligence is Difficult
(01:16:45) The Lighter Side
The original text contained 23 images which were described by AI.
---
First published:
February 27th, 2025Source:
---
https://www.lesswrong.com/posts/v5dpeuj4qPxngcb4d/ai-105-hey-there-alexaNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
Anthropic has reemerged from stealth and offers us Claude 3.7.
Given this is named Claude 3.7, an excellent choice, from now on this blog will refer to what they officially call Claude Sonnet 3.5 (new) as Sonnet 3.6.
Claude 3.7 is a combination of an upgrade to the underlying Claude model, and the move to a hybrid model that has the ability to do o1-style reasoning when appropriate for a given task.
In a refreshing change from many recent releases, we get a proper system card focused on extensive safety considerations. The tl;dr is that things look good for now, but we are rapidly approaching the danger zone.
The cost for Sonnet 3.7 via the API is the same as it was for 3.6, $5/$15 for million. If you use extended thinking, you have to pay for the thinking tokens.
They also introduced a [...]
---
Outline:
(01:17) Executive Summary
(03:09) Part 1: Capabilities
(03:14) Extended Thinking
(04:17) Claude Code
(06:52) Data Use
(07:11) Benchmarks
(08:25) Claude Plays Pokemon
(09:21) Private Benchmarks
(16:14) Early Janus Takes
(18:31) System Prompt
(24:25) Easter Egg
(25:50) Vibe Coding Reports
(32:53) Practical Coding Advice
(35:02) The Future
(36:05) Part 2: Safety and the System Card
(36:24) Claude 3.7 Tested as ASL-2
(38:15) The RSP Evaluations That Concluded Claude 3.7 is ASL-2
(40:41) ASL-3 is Coming Soon, and With That Comes Actual Risk
(43:31) Reducing Unnecessary Refusals
(45:11) Mundane Harm Evolutions
(45:53) Risks From Computer Use
(47:15) Chain of Thought Faithfulness
(48:53) Alignment Was Not Faked
(49:38) Excessive Focus on Passing Tests
(51:13) The Lighter Side
The original text contained 25 images which were described by AI.
---
First published:
February 26th, 2025Source:
---
https://www.lesswrong.com/posts/Wewdcd52zwfdGYqAi/time-to-welcome-claude-3-7Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
This is a post in two parts.
The first half is the post is about Grok's capabilities, now that we’ve all had more time to play around with it. Grok is not as smart as one might hope and has other issues, but it is better than I expected and for now has its place in the rotation, especially for when you want its Twitter integration.
That was what this post was supposed to be about.
Then the weekend happened, and now there's also a second half. The second half is about how Grok turned out rather woke and extremely anti-Trump and anti-Musk, as well as trivial to jailbreak, and the rather blunt things xAI tried to do about that. There was some good transparency in places, to their credit, but a lot of trust has been lost. It will be extremely difficult to win it [...]
---
Outline:
(01:21) Zvi Groks Grok
(03:39) Grok the Cost
(04:29) Grok the Benchmark
(06:02) Fun with Grok
(08:33) Others Grok Grok
(11:26) Apps at Play
(12:38) Twitter Groks Grok
(13:38) Grok the Woke
(19:06) Grok is Misaligned
(20:07) Grok Will Tell You Anything
(24:29) xAI Keeps Digging (1)
(29:21) xAI Keeps Digging (2)
(39:14) What the Grok Happened
(43:29) The Lighter Side
The original text contained 24 images which were described by AI.
---
First published:
February 24th, 2025Source:
---
https://www.lesswrong.com/posts/tpLfqJhxcijf5h23C/grok-grokNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
While we wait for the verdict on Anthropic's Claude Sonnet 3.7, today seems like a good day to catch up on the queue and look at various economics-related things.
Table of Contents
The Trump Tax Proposals.Taxing Unrealized Capital Gains.Extremely High Marginal Tax Rates.Trade Barriers By Any Name Are Terrible.Destroying People's Access to Credit.Living Paycheck to Paycheck.Oh California.Chinese Venture Capital Death Spiral.There is Someone Elon Musk Forgot to Ask.Should Have Gone With the Sports Almanac.Are You Better Off Than You Were Right Before the Election?.Are You Better Off Than You Were Before the Price Level Rose?.Most People Have No Idea How Insurance Works.Do Not Spend Too Much Attention on Your Investments.Preferences About Insider Training are Weird.I [...]---
Outline:
(00:15) The Trump Tax Proposals
(02:36) Taxing Unrealized Capital Gains
(03:00) Extremely High Marginal Tax Rates
(05:24) Trade Barriers By Any Name Are Terrible
(06:11) Destroying People's Access to Credit
(06:35) Living Paycheck to Paycheck
(09:54) Oh California
(10:12) Chinese Venture Capital Death Spiral
(11:20) There is Someone Elon Musk Forgot to Ask
(14:39) Should Have Gone With the Sports Almanac
(17:58) Are You Better Off Than You Were Right Before the Election?
(18:13) Are You Better Off Than You Were Before the Price Level Rose?
(25:43) Most People Have No Idea How Insurance Works
(27:25) Do Not Spend Too Much Attention on Your Investments
(28:43) Preferences About Insider Training are Weird
(30:12) I Will Not Allocate Scarce Resources Via Price
(30:33) Minimum Wages, Employment and the Equilibrium
(31:22) The National Debt
(36:09) In Brief
The original text contained 1 footnote which was omitted from this narration.
The original text contained 6 images which were described by AI.
---
First published:
February 25th, 2025Source:
---
https://www.lesswrong.com/posts/AAKXjRmBRbJJwGthT/economics-roundup-5Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
OpenAI made major revisions to their Model Spec.
It seems very important to get this right, so I’m going into the weeds.
This post thus gets farther into the weeds than most people need to go. I recommend most of you read at most the sections of Part 1 that interest you, and skip Part 2.
I looked at the first version last year. I praised it as a solid first attempt.
Table of Contents
Part 1Conceptual Overview.Change Log.Summary of the Key Rules.Three Goals.Three Risks.The Chain of Command.The Letter and the Spirit.Part 2Stay in Bounds: Platform Rules.The Only Developer Rule.Mental Health.What is on the Agenda.Liar Liar.Still Kind of a Liar Liar.Well, Yes [...]---
Outline:
(00:30) Part 1
(00:33) Conceptual Overview
(05:51) Change Log
(07:25) Summary of the Key Rules
(11:49) Three Goals
(15:51) Three Risks
(20:07) The Chain of Command
(26:14) The Letter and the Spirit
(29:30) Part 2
(29:33) Stay in Bounds: Platform Rules
(47:19) The Only Developer Rule
(49:19) Mental Health
(50:38) What is on the Agenda
(56:35) Liar Liar
(01:01:56) Still Kind of a Liar Liar
(01:07:42) Well, Yes, Okay, Sure
(01:10:14) I Am a Good Nice Bot
(01:20:55) A Conscious Choice
(01:21:49) Part 3
(01:21:52) The Super Secret Instructions
(01:24:45) The Super Secret Model Spec Details
(01:27:43) A Final Note
The original text contained 29 images which were described by AI.
---
First published:
February 21st, 2025Source:
---
https://www.lesswrong.com/posts/ntQYby9G8A85cEeY6/on-openai-s-model-spec-2-0Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
The Trump Administration is on the verge of firing all ‘probationary’ employees in NIST, as they have done in many other places and departments, seemingly purely because they want to find people they can fire. But if you fire all the new employees and recently promoted employees (which is that ‘probationary’ means here) you end up firing quite a lot of the people who know about AI or give the government state capacity in AI.
This would gut not only America's AISI, its primary source of a wide variety of forms of state capacity and the only way we can have insight into what is happening or test for safety on matters involving classified information. It would also gut our ability to do a wide variety of other things, such as reinvigorating American semiconductor manufacturing. It would be a massive own goal for the United States, on every [...]
---
Outline:
(01:14) Language Models Offer Mundane Utility
(05:44) Language Models Don't Offer Mundane Utility
(10:13) Rug Pull
(12:19) We're In Deep Research
(21:12) Huh, Upgrades
(30:28) Seeking Deeply
(35:26) Fun With Multimedia Generation
(35:41) The Art of the Jailbreak
(36:26) Get Involved
(37:09) Thinking Machines
(41:13) Introducing
(42:58) Show Me the Money
(44:55) In Other AI News
(53:31) By Any Other Name
(56:06) Quiet Speculations
(59:37) The Copium Department
(01:02:33) Firing All 'Probationary' Federal Employees Is Completely Insane
(01:10:28) The Quest for Sane Regulations
(01:12:18) Pick Up the Phone
(01:14:24) The Week in Audio
(01:16:19) Rhetorical Innovation
(01:18:50) People Really Dislike AI
(01:20:45) Aligning a Smarter Than Human Intelligence is Difficult
(01:22:34) People Are Worried About AI Killing Everyone
(01:23:51) Other People Are Not As Worried About AI Killing Everyone
(01:24:16) The Lighter Side
The original text contained 17 images which were described by AI.
---
First published:
February 20th, 2025Source:
---
https://www.lesswrong.com/posts/bozSPnkCzXBjDpbHj/ai-104-american-state-capacity-on-the-brinkNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
That title is Elon Musk's fault, not mine, I mean, sorry not sorry:
Table of Contents
Release the Hounds.The Expectations Game.Man in the Arena.The Official Benchmarks.The Inevitable Pliny.Heart in the Wrong Place.Where Is Your Head At.Individual Reactions.Grok on Grok.Release the Hounds
Grok 3 is out. It mostly seems like no one cares.
I expected this, but that was because I expected Grok 3 to not be worth caring about.
Instead, no one cares for other reasons, like the rollout process being so slow (in a poll on my Twitter this afternoon, the vast majority of people hadn’t used it) and access issues and everyone being numb to another similar model and the pace of events. And because everyone is so sick of the hype.
[...]---
Outline:
(00:36) Release the Hounds
(02:11) The Expectations Game
(06:45) Man in the Arena
(07:29) The Official Benchmarks
(09:35) The Inevitable Pliny
(12:01) Heart in the Wrong Place
(14:16) Where Is Your Head At
(15:10) Individual Reactions
(28:39) Grok on Grok
The original text contained 13 images which were described by AI.
---
First published:
February 19th, 2025Source:
---
https://www.lesswrong.com/posts/WNYvFCkhZvnwAPzJY/go-grok-yourselfNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
It seems like as other things drew our attention more, medical news slowed down. The actual developments, I have no doubt, are instead speeding up – because AI.
Note that this post intentionally does not cover anything related to the new Administration, or its policies.
Table of Contents
Some People Need Practical Advice.Good News, Everyone.Bad News.Life Extension.Doctor Lies to Patient.Study Lies to Public With Statistics.Area Man Discovers Information Top Doctors Missed.Psychiatric Drug Prescription.H5N1.WHO Delenda Est.Medical Ethicists Take Bold Anti-Medicine Stance.Rewarding Drug Development.Not Rewarding Device Developers.Addiction.Our Health Insurance Markets are Broken.Some People Need Practical Advice
If you ever have to go to the hospital for any reason, suit up, or at least look [...]
---
Outline:
(00:22) Some People Need Practical Advice
(00:32) Good News, Everyone
(03:13) Bad News
(04:05) Life Extension
(04:42) Doctor Lies to Patient
(06:36) Study Lies to Public With Statistics
(08:46) Area Man Discovers Information Top Doctors Missed
(10:49) Psychiatric Drug Prescription
(11:30) H5N1
(12:43) WHO Delenda Est
(13:02) Medical Ethicists Take Bold Anti-Medicine Stance
(13:41) Rewarding Drug Development
(16:19) Not Rewarding Device Developers
(17:59) Addiction
(18:27) Our Health Insurance Markets are Broken
The original text contained 4 images which were described by AI.
---
First published:
February 18th, 2025Source:
---
https://www.lesswrong.com/posts/nSuYdFzdNA7rrzmyJ/medical-roundup-4Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
I have been debating how to cover the non-AI aspects of the Trump administration, including the various machinations of DOGE. I felt it necessary to have an associated section this month, but I have attempted to keep such coverage to a minimum, and will continue to do so. There are too many other things going on, and plenty of others are covering the situation.
Table of Contents
Bad News.Antisocial Media.Variously Effective Altruism.The Forbidden Art of Fundraising.There Was Ziz Thing.That's Not Very Nice.The Unbearable Weight Of Lacking Talent.How to Have More Agency.Government Working: Trump Administration Edition. Government Working.The Boolean Illusion.Nobody Wants This.We Technically Didn’t Start the Fire.Good News, Everyone.A Well Deserved Break.Opportunity Knocks. -
This post covers three recent shenanigans involving OpenAI.
In each of them, OpenAI or Sam Altman attempt to hide the central thing going on.
First, in Three Observations, Sam Altman's essay pitches our glorious AI future while attempting to pretend the downsides and dangers don’t exist in some places, and in others admitting we’re not going to like those downsides and dangers but he's not about to let that stop him. He's going to transform the world whether we like it or not.
Second, we have Frog and Toad, or There Is No Plan, where OpenAI reveals that its plan for ensuring AIs complement humans rather than AIs substituting for humans is to treat this as a ‘design choice.’ They can simply not design AIs that will be substitutes. Except of course this is Obvious Nonsense in context, with all the talk of remote workers, and [...]
---
Outline:
(01:52) Three Observations
(11:38) Frog and Toad (or There Is No Plan)
(18:45) A Trade Offer Has Arrived
The original text contained 5 images which were described by AI.
---
First published:
February 14th, 2025Source:
---
https://www.lesswrong.com/posts/drHsruvnkCYweMJp7/the-mask-comes-off-a-trio-of-talesNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
The main event this week was the disastrous Paris AI Anti-Safety Summit. Not only did we not build upon the promise of the Bletchley and Seoul Summits, the French and Americans did their best to actively destroy what hope remained, transforming the event into a push for a mix of nationalist jingoism, accelerationism and anarchism. It's vital and also difficult not to panic or despair, but it doesn’t look good.Another major twist was that Elon Musk made a $97 billion bid for OpenAI's nonprofit arm and its profit and control interests in OpenAI's for-profit arm. This is a serious complication for Sam Altman's attempt to buy those same assets for $40 billion, in what I’ve described as potentially the largest theft in human history. I’ll be dealing with that tomorrow, along with two other developments in my ongoing OpenAI series The Mask Comes Off. In Altman's Three Observations, he [...]
---
Outline:
(01:55) Language Models Offer Mundane Utility
(06:03) Language Models Don't Offer Mundane Utility
(08:07) We're in Deep Research
(13:54) Huh, Upgrades
(20:56) Seeking Deeply
(24:25) Smooth Operator
(29:15) They Took Our Jobs
(33:34) Maxwell Tabarrok Responds on Future Wages
(41:56) The Art of the Jailbreak
(46:34) Get Involved
(48:45) Introducing
(51:17) Show Me the Money
(53:20) In Other AI News
(56:12) Quiet Speculations
(01:02:05) The Quest for Sane Regulations
(01:04:53) The Week in Audio
(01:06:40) The Mask Comes Off
(01:08:20) Rhetorical Innovation
(01:21:35) Getting Tired of Winning
(01:24:36) People Really Dislike AI
(01:25:47) Aligning a Smarter Than Human Intelligence is Difficult
(01:27:20) Sufficiently Capable AIs Effectively Acquire Convergent Utility Functions
(01:36:29) People Are Worried About AI Killing Everyone
(01:47:03) Other People Are Not As Worried About AI Killing Everyone
(01:50:38) The Lighter Side
The original text contained 18 images which were described by AI.
---
First published:
February 13th, 2025Source:
---
https://www.lesswrong.com/posts/Lmqi4x5zntjSxfdPg/ai-103-show-me-the-moneyNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
It doesn’t look good.
What used to be the AI Safety Summits were perhaps the most promising thing happening towards international coordination for AI Safety.
This one was centrally coordination against AI Safety.
In November 2023, the UK Bletchley Summit on AI Safety set out to let nations coordinate in the hopes that AI might not kill everyone. China was there, too, and included.
The practical focus was on Responsible Scaling Policies (RSPs), where commitments were secured from the major labs, and laying the foundations for new institutions.
The summit ended with The Bletchley Declaration (full text included at link), signed by all key parties. It was the usual diplomatic drek, as is typically the case for such things, but it centrally said there are risks, and so we will develop policies to deal with those risks.
And it ended with a commitment [...]
---
Outline:
(02:03) An Actively Terrible Summit Statement
(05:45) The Suicidal Accelerationist Speech by JD Vance
(14:37) What Did France Care About?
(17:12) Something To Remember You By: Get Your Safety Frameworks
(24:05) What Do We Think About Voluntary Commitments?
(27:29) This Is the End
(36:18) The Odds Are Against Us and the Situation is Grim
(39:52) Don't Panic But Also Face Reality
The original text contained 4 images which were described by AI.
---
First published:
February 12th, 2025Source:
---
https://www.lesswrong.com/posts/qYPHryHTNiJ2y6Fhi/the-paris-ai-anti-safety-summitNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
Not too long ago, OpenAI presented a paper on their new strategy of Deliberative Alignment.
The way this works is that they tell the model what its policies are and then have the model think about whether it should comply with a request.
This is an important transition, so this post will go over my perspective on the new strategy.
Note the similarities, and also differences, with Anthropic's Constitutional AI.
How Deliberative Alignment Works
We introduce deliberative alignment, a training paradigm that directly teaches reasoning LLMs the text of human-written and interpretable safety specifications, and trains them to reason explicitly about these specifications before answering.
We used deliberative alignment to align OpenAI's o-series models, enabling them to use chain-of-thought (CoT) reasoning to reflect on user prompts, identify relevant text from OpenAI's internal policies, and draft safer responses.
Our approach achieves highly precise [...]
---
Outline:
(00:29) How Deliberative Alignment Works
(03:27) Why This Worries Me
(07:49) For Mundane Safety It Works Well
The original text contained 3 images which were described by AI.
---
First published:
February 11th, 2025Source:
---
https://www.lesswrong.com/posts/CJ4yywLBkdRALc4sT/on-deliberative-alignmentNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
Scott Alexander famously warned us to Beware Trivial Inconveniences.
When you make a thing easy to do, people often do vastly more of it.
When you put up barriers, even highly solvable ones, people often do vastly less.
Let us take this seriously, and carefully choose what inconveniences to put where.
Let us also take seriously that when AI or other things reduce frictions, or change the relative severity of frictions, various things might break or require adjustment.
This applies to all system design, and especially to legal and regulatory questions.
Table of Contents
Levels of Friction (and Legality).Important Friction Principles.Principle #1: By Default Friction is Bad.Principle #3: Friction Can Be Load Bearing.Insufficient Friction On Antisocial Behaviors Eventually Snowballs.Principle #4: The Best Frictions Are Non-Destructive.Principle #8: The Abundance [...]---
Outline:
(00:40) Levels of Friction (and Legality)
(02:24) Important Friction Principles
(05:01) Principle #1: By Default Friction is Bad
(05:23) Principle #3: Friction Can Be Load Bearing
(07:09) Insufficient Friction On Antisocial Behaviors Eventually Snowballs
(08:33) Principle #4: The Best Frictions Are Non-Destructive
(09:01) Principle #8: The Abundance Agenda and Deregulation as Category 1-ification
(10:55) Principle #10: Ensure Antisocial Activities Have Higher Friction
(11:51) Sports Gambling as Motivating Example of Necessary 2-ness
(13:24) On Principle #13: Law Abiding Citizen
(14:39) Mundane AI as 2-breaker and Friction Reducer
(20:13) What To Do About All This
The original text contained 1 image which was described by AI.
---
First published:
February 10th, 2025Source:
---
https://www.lesswrong.com/posts/xcMngBervaSCgL9cu/levels-of-frictionNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
-
This week we got a revision of DeepMind's safety framework, and the first version of Meta's framework. This post covers both of them.
Table of Contents
Meta's RSP (Frontier AI Framework).DeepMind Updates its Frontier Safety Framework.What About Risk Governance.Where Do We Go From Here?Here are links for previous coverage of: DeepMind's Framework 1.0, OpenAI's Framework and Anthropic's Framework.
Meta's RSP (Frontier AI Framework)
Since there is a law saying no two companies can call these documents by the same name, Meta is here to offer us its Frontier AI Framework, explaining how Meta is going to keep us safe while deploying frontier AI systems.
I will say up front, if it sounds like I’m not giving Meta the benefit of the doubt here, it's because I am absolutely not giving Meta the benefit of [...]
---
Outline:
(00:14) Meta's RSP (Frontier AI Framework)
(16:10) DeepMind Updates its Frontier Safety Framework
(31:05) What About Risk Governance
(33:42) Where Do We Go From Here?
The original text contained 12 images which were described by AI.
---
First published:
February 7th, 2025Source:
---
https://www.lesswrong.com/posts/etqbEF4yWoGBEaPro/on-the-meta-and-deepmind-safety-frameworksNarrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
- Visa fler