I liked this talk.
https://youtu.be/WWoyWNhx2XU?si=g0lkoY7gAU-bYRjd
Benjamin Mann is a co-founder of Anthropic. Prior to Anthropic, Ben was one of the architects of GPT-3 at OpenAI. He left OpenAI driven by the mission to ensure that AI benefits humanity. In this talk, Ben opens up about the accelerating progress in AI and the urgent need to steer it responsibly.
In this conversation, we discuss:
1. The inside story of leaving OpenAI with the entire safety team to start Anthropic
2. How Metaβs $100M offers reveal the true market price of top AI talent
3. Why AI progress is still accelerating (not plateauing), and how most people misjudge the exponential
4. Benβs βeconomic Turing testβ for knowing when weβve achieved AGIβand why itβs likely coming by 2027-2028
5. Why he believes 20% unemployment is inevitable
6. The AI nightmare scenarios that concern him mostβand how he believes we can still avoid them
7. How focusing on AI safety created Claudeβs beloved personality
8. What three skills heβs teaching his kids instead of traditional academics
https://youtu.be/WWoyWNhx2XU?si=g0lkoY7gAU-bYRjd
YouTube
Anthropic co-founder: AGI predictions, leaving OpenAI, what keeps him up at night | Ben Mann
Benjamin Mann is a co-founder of Anthropic, an AI startup dedicated to building aligned, safety-first AI systems. Prior to Anthropic, Ben was one of the architects of GPT-3 at OpenAI. He left OpenAI driven by the mission to ensure that AI benefits humanity.β¦
π₯15β€6π6
Even OpenAI engineers use Claude.
Well, they wanted to.
https://www.wired.com/story/anthropic-revokes-openais-access-to-claude/
Well, they wanted to.
https://www.wired.com/story/anthropic-revokes-openais-access-to-claude/
WIRED
Anthropic Revokes OpenAI's Access to Claude
OpenAI lost access to the Claude API this week after Anthropic claimed the company was violating its terms of service.
π€£21π€―4β€3π2
Just some simple thoughts on jobs.
Some people believe there will be no work/jobs in the post-AGI world. I don't think so even though I consider myself AGI-pilled. Mostly because I'm cynical about human nature.
I cannot imagine a world without any problems. Look at the current state, there are wars, climate change, drug epidemic, idiotic country leaders, poor education. That will take a long time to fix. As long as there are problems, there will be work to do.
Second, it is true that work != job. I believe there will always be some scarcity of something, and in particular I don't expect the distance between poor and rich (or low status vs high status) to diminish any time soon. I don't expect a human will be satisfied, ever - there is always more to desire. Earning that will require work in exchange for what the person wants, or perhaps something that can get them what they want (aka money, or some other form of credit). That's a job.
And yes, it is quite possible that those jobs will be much easier, to the extent that you might not consider them jobs at all (is game streaming a job?). But I'm sure most people from 1825 wouldn't consider what we do jobs either.
The transitional period is gonna be rough though, and that's probably the toughest part. People who don't want to change will be hit the most.
Some people believe there will be no work/jobs in the post-AGI world. I don't think so even though I consider myself AGI-pilled. Mostly because I'm cynical about human nature.
I cannot imagine a world without any problems. Look at the current state, there are wars, climate change, drug epidemic, idiotic country leaders, poor education. That will take a long time to fix. As long as there are problems, there will be work to do.
Second, it is true that work != job. I believe there will always be some scarcity of something, and in particular I don't expect the distance between poor and rich (or low status vs high status) to diminish any time soon. I don't expect a human will be satisfied, ever - there is always more to desire. Earning that will require work in exchange for what the person wants, or perhaps something that can get them what they want (aka money, or some other form of credit). That's a job.
And yes, it is quite possible that those jobs will be much easier, to the extent that you might not consider them jobs at all (is game streaming a job?). But I'm sure most people from 1825 wouldn't consider what we do jobs either.
The transitional period is gonna be rough though, and that's probably the toughest part. People who don't want to change will be hit the most.
π25π₯12β€6
Does this count as new knowledge? No human knew about these vulnerabilities and now we do
Knowledge discovery is a search problem.
https://x.com/argvee/status/1952390039700431184?s=46
Knowledge discovery is a search problem.
Today as part of our commitment to transparency in this space, we are proud to announce that we have reported the first 20 vulnerabilities discovered using our AI-based "Big Sleep" system powered by Gemini
https://x.com/argvee/status/1952390039700431184?s=46
π6π€‘3
5x increase of max context length.
Note: tokens over 200k are more expensive if you use raw API
https://www.anthropic.com/news/1m-context
Note: tokens over 200k are more expensive if you use raw API
https://www.anthropic.com/news/1m-context
Claude
Claude Sonnet 4 now supports 1M tokens of context | Claude
Claude Sonnet 4 now supports up to 1 million tokens of contextβa 5x increase that lets you process entire codebases, synthesize extensive document sets, and build agents that maintain coherence across hundreds of tool calls.
π₯25π7β€2π1
living in uncertainty
each of us is wrong sometimes. when a prediction you made turns out to be false, it's a good exercise to back-propagate it to your world model (what did i miss?) and then do a forward pass to understand what other implications this discovery has (what else must be true that i thought is false). usually it's not a big deal. often it is some wrong assumption/bias which impacts few things. you update your world model and move on
the ai of today was in the realm of fantasy a few years ago. i tried to find what we missed, but all plausible theories explaining llms suggest that we missed something very fundamental, and which has a lot of profound implications. among all world model updates i consider, the delta is large (the rabbit hole is deep), but also i'm not certain which one is the most accurate. just as an example, i am not aware of a theory that would explain why llms are getting good at coding but wont explain why they will eventually become good at everything else.
moreover it is dynamic. last year the talk was about prompt engineering. this year it's agents. i expect it to be different next year, and in general i expect things to continue to evolve. more capabilities will be developed, more work automated and more population waking up
it can be challenging to live like that: we like certainty. you might be used to have some life plan that you follow (say you are a prospective student who chooses your major). any plan has to assume that things that were true during planning will stay true. ok, maybe instead of a static plan you have a more dynamic strategy, essentially a function that accepts the current state and returns suggested actions. well, that function is still compressed knowledge of what works and what doesn't, but that knowledge might also get stale as the world changes
certainty is a luxury. i expect it to continue decreasing. i'm not even sure it will be coming back in our lifetimes, so I'd suggest to start getting used to a world where you can't be very confident in the increasing number of things (this is particularly hard for very smart people who are right most of the time and got used to that)
the only advice i have is gain some humility, be open minded and get ready to constantly adapt to the world changing under our feet
each of us is wrong sometimes. when a prediction you made turns out to be false, it's a good exercise to back-propagate it to your world model (what did i miss?) and then do a forward pass to understand what other implications this discovery has (what else must be true that i thought is false). usually it's not a big deal. often it is some wrong assumption/bias which impacts few things. you update your world model and move on
the ai of today was in the realm of fantasy a few years ago. i tried to find what we missed, but all plausible theories explaining llms suggest that we missed something very fundamental, and which has a lot of profound implications. among all world model updates i consider, the delta is large (the rabbit hole is deep), but also i'm not certain which one is the most accurate. just as an example, i am not aware of a theory that would explain why llms are getting good at coding but wont explain why they will eventually become good at everything else.
moreover it is dynamic. last year the talk was about prompt engineering. this year it's agents. i expect it to be different next year, and in general i expect things to continue to evolve. more capabilities will be developed, more work automated and more population waking up
it can be challenging to live like that: we like certainty. you might be used to have some life plan that you follow (say you are a prospective student who chooses your major). any plan has to assume that things that were true during planning will stay true. ok, maybe instead of a static plan you have a more dynamic strategy, essentially a function that accepts the current state and returns suggested actions. well, that function is still compressed knowledge of what works and what doesn't, but that knowledge might also get stale as the world changes
certainty is a luxury. i expect it to continue decreasing. i'm not even sure it will be coming back in our lifetimes, so I'd suggest to start getting used to a world where you can't be very confident in the increasing number of things (this is particularly hard for very smart people who are right most of the time and got used to that)
the only advice i have is gain some humility, be open minded and get ready to constantly adapt to the world changing under our feet
π26β€βπ₯11π₯6β€2π2π1π³1
I appreciate being mentioned in top-30 AI people in Uzbekistan. Thanks.
https://yuksalish.org/uz/news_detail/835
https://yuksalish.org/uz/news_detail/835
π₯55π14β€8π6π€ͺ1
don't let AIs help you spiral into your craziness. if you are vulnerable to that, don't use sycophantic AIs
https://thezvi.wordpress.com/2025/09/16/ai-craziness-notes/
https://thezvi.wordpress.com/2025/09/16/ai-craziness-notes/
π11π»3
Anthropic's first postmortem about the model being dumber, with a level of detail that we usually don't share
https://www.anthropic.com/engineering/a-postmortem-of-three-recent-issues
https://www.anthropic.com/engineering/a-postmortem-of-three-recent-issues
Anthropic
A postmortem of three recent issues
This is a technical report on three bugs that intermittently degraded responses from Claude. Below we explain what happened, why it took time to fix, and what we're changing.
π₯16π’2
what an idiot. i mean, the fact that he is an idiot is not new, but this is the new level
https://x.com/whitehouse/status/1969147079478989220?s=46
https://x.com/whitehouse/status/1969147079478989220?s=46
X (formerly Twitter)
The White House (@WhiteHouse) on X
π° NEW @Bloomberg: Trump to Add New $100,000 Fee for H-1B Visas in Latest Crackdown.
π±11π―8π€―7π2π’2β€1π1
OpenAI released a new eval that measures performance on economically valuable, real-world tasks across 44 occupations.
https://openai.com/index/gdpval/
https://openai.com/index/gdpval/
π₯18β€3π3
great example of why i didn't even consider applying for openai. think of implications in a country with a racist president
https://x.com/gabrielpeterss4/status/1973120058907041902?s=46
https://x.com/gabrielpeterss4/status/1973120058907041902?s=46
X (formerly Twitter)
gabriel (@gabriel1) on X
i have the most liked video on sora 2 right now, i will be enjoying this short moment while it lasts
cctv footage of sam stealing gpus at target for sora inference
cctv footage of sam stealing gpus at target for sora inference
π21β€6π1π1π1
You can now connect Slack to Claude. It can search your workspace channels, DMs, and files/gdocs to provide context for deep work.
You can also connect Claude app to slack, e.g. ask something in the app and claude can read your slack, search info there, etc.
Video below
https://x.com/claudeai/status/1973445694305468597?s=46
You can also connect Claude app to slack, e.g. ask something in the app and claude can read your slack, search info there, etc.
Video below
https://x.com/claudeai/status/1973445694305468597?s=46
X (formerly Twitter)
Claude (@claudeai) on X
Claude is now available in Slack.
Chat with Claude through DMs, tag @.Claude in threads, or use the AI assistant panelβwith access to web search, document analysis, and your connected tools.
Chat with Claude through DMs, tag @.Claude in threads, or use the AI assistant panelβwith access to web search, document analysis, and your connected tools.
π9π₯8β€5
π¦ i recommend spending a year with Rust
i don't think i can explain all the reasons why do that in a way that's both short and clear. most likely i'll lose the reader in the middle of the post before i'd get to the point. it is only after some first-hand prolonged experience of learning the Rust way you start getting it.
just trust me on this π go ahead and do yourself a favor
fair warning: first 6mo can be painful, but we have LLMs now that help a lot
i don't think i can explain all the reasons why do that in a way that's both short and clear. most likely i'll lose the reader in the middle of the post before i'd get to the point. it is only after some first-hand prolonged experience of learning the Rust way you start getting it.
just trust me on this π go ahead and do yourself a favor
fair warning: first 6mo can be painful, but we have LLMs now that help a lot
π«‘42β€11π₯5π4π4π2π2
haiku 4.5 (just released) is as smart as sonnet 4.0, but it's 2x faster and 3x cheaper. i've been using it in claude code for a while (primarily because of speed) and i can recommend it. i use it more often than sonnet 4.5 and definitely more than opus
https://www.anthropic.com/news/claude-haiku-4-5
https://www.anthropic.com/news/claude-haiku-4-5
π23β€10π₯8
Addressing seemingly common misunderstanding.
- Sonnet 4.5 is smarter than Opus 4.1.
- Haiku 4.5 nearly as smart than Sonnet 4.0
how come? Scaling laws suggest that the intelligence of models grows with scale (aka the bitter lesson). We increase training scale all the time, so it is not surprising that a newer model is more intelligent than an older model.
Besides, smaller models are:
- much faster, so you are getting more done
- cheaper, so your quota lasts longer
- Sonnet 4.5 is smarter than Opus 4.1.
- Haiku 4.5 nearly as smart than Sonnet 4.0
how come? Scaling laws suggest that the intelligence of models grows with scale (aka the bitter lesson). We increase training scale all the time, so it is not surprising that a newer model is more intelligent than an older model.
Besides, smaller models are:
- much faster, so you are getting more done
- cheaper, so your quota lasts longer
π₯18π7β€2πΎ2