Amodei does not mean that things are plateauing (i.e. the exponential will no longer hold), but rather uses "end" closer to the notion of "endgame," that is we are getting to the point where all benchmarks pegged to human ability will be saturated and the AI systems will be better than any human at any cognitive task.
Amodei lays this out here:
> [with regards to] the “country of geniuses in a data center”. My picture for that, if you made me guess, is one to two years, maybe one to three years. It’s really hard to tell. I have a strong view—99%, 95%—that all this will happen in 10 years. I think that’s just a super safe bet. I have a hunch—this is more like a 50/50 thing—that it’s going to be more like one to two [years], maybe more like one to three.
This is why Amodei opens with
> What has been the most surprising thing is the lack of public recognition of how close we are to the end of the exponential. To me, it is absolutely wild that you have people — within the bubble and outside the bubble — talking about the same tired, old hot-button political issues, when we are near the end of the exponential.
Whether you agree with him is of course a different matter altogether, but a clearer phrasing would probably be "We are near the endgame."
It is a 2+hrs video and hence a summary of main themes is welcome.
Nothing that I have seen described here on HN or elsewhere, by the most enthusiast users of AI, who claim that their own productivity has been multiplied, does not demonstrate performance in cognitive tasks even remotely comparable with that of a competent human, much less better performance.
All that I see is that the AI systems outperform humans for various tasks only because they had access in their training data to much more information than most humans are allowed to access, because they do not have enough money to obtain such access, both because the various copyright paywalls and also because of the actual cost of storage and retrieval systems.
Using an AI agent may be faster than if you were given access to the training data and you would use conventional search tools on it, but the speed may be illusory, because when I search something and I have access to the original sources I can validate the search results faster and with much more certainty than when I try to ponder about the correctness of what AI has provided, e.g. whether a program produced by it really does what I have requested and it is bug free (in comparison with having access to its training programs and being able to choose myself what to copy and paste).
I hope that paid access to AI tools gives better results, but the AI replies that popular search engines, like Google and Bing, force upon their users have made Internet searches much worse not better, as their answers always contain something else than I want, and this is in the best case, when the answers are not plainly wrong.
You should get yourself a paid subscription. Honest advice. The difference between agentic workflow vs single-shot questions in free-tier services is night and day. Building context and letting the model have access to your code is the largest differentiator between "wtf, I don't need this" and "wtf".
> All that I see is that the AI systems outperform humans for various tasks only because they had access in their training data to much more information than most humans are allowed to access, because they do not have enough money to obtain such access
Humans cannot even theoretically read and consume the volume of data the models can do so it's not really about the money - it's more about the infinite amount of time humans would need to have and the extremely large cognitive load it would impose on them. How many people can even synthesize so much diverse topics at high and constant pace? None or very little.
Also, models are proven to generalize very well so having access to your codebase during the training phase is not necessary for them to provide you with the correct answers. Give it a try.
Models being able to generalize very well is one of the ways AI labs think they may reach the "AI systems will be better than any human at any cognitive task" goal. I am not convinced that this will be the only sauce needed but I am also not too skeptic about it too, given the speed at which AI capabilities unfolded, especially during the 2025.
I think we already reached the point where it's safe to say that "AI systems are better than many humans at most cognitive tasks". I can see it myself on the project I am currently working on. These are not the top-tier developers. And when I talk to the top-tier ones I have previously worked with, we share the similar sentiment. The only difference might be "AI systems are much faster than many humans at most congitive tasks".
I give it a question (narrow or really broad), and the model does a bunch of web searches using subagents, to try and get a comprehensive answer using current results
The important part is, when the model answers, I have it cite its sources, using direct links. So, I can directly confirm the accuracy and quality of any info it finds
It's been super helpful. I can give it super broad questions like "Here's the architecture and environment details I'm planning for a new project. Can you see if there's any known issues with this setup". Then, it'll give me direct links + summaries to any relevant pages.
Saves a ton of time manually searching through the haystack, and so far, the latest models are pretty good about not missing important things (and catching plenty of things I missed)
Unsurprisingly, we were able to build a demo platform within a few days. But when we started building the actual platform, we realized that the code generated by Claude is hard to extend, and a lot of replanning and reworking needs to be done every time you try to add a major feature.
This brought our confidence level down. We still want to believe that Claude will help in generating code. But I no longer believe that Claude will be able to write complex software on its own.
Now we are treating Claude as a junior person on the team and give it well-defined, specific tasks to complete.
Usually the verification and testing is the most time consuming part.
I am working on graphical application like AutoCAD, for the context.
If they can run a tool from the terminal, see all the output in text format, and have a clear 'success' criteria, then they're usually able to figure out the issue and fix it (often with spaghetti code patching, but it does at least fix the bug)
I think the testing/verification part is going to keep getting better, as we figure out better tools the AI can use here (ex, parsing the accessibility tree in a web UI to click around in it and verify)
Seriously?
Between a human and a malformed humunculus of piddling intelligence?
On the one hand, there is a lot of hype, an incredible amount, actually, but on the other, we have been observing in real time a technological miracle that gets better by the week.
We have no idea what, five years from now, the coding agent will be able to develop.
We have successfully put Claude in huge multi-thousands pr long with projects.
But this meant that:
1. Solid architectural and design decisions were made already after much trial and error
2. They were further refined and refactored
3. Countless hours have been spent in documenting, writing proper skills and architectural and best practice documents
Only then Claude started paying off, and even then it's an iterative process where you need to understand why it tries to hack his way out, etc, what to check, what to supervise.
Seriously if you think you can just Claude create some project..
Just fork an existing one that does some larger % of what you need and spend most of the initial time scaffolding it to be ai friendly.
Also, you need to invest in harnessing, giving tools and ways to the LLM to not go off rails.
Strongly typed languages, plenty of compilation and diagnostics tools, access to debuggers or browser mcps, etc.
It's not impossible, but you need to approach it with an experimentation approach, not drinking Kool aid.
The idea of AI being able to "code" is that it is able to do all this planning and architectural work. It cant. But its sold as though it is. Thats where the bubble is
"on its own" is doing a lot of work here. Dario went into the differences in this very podcast: "Most code is written by agents" is not the same as "most code is written without or independent of human input".
I suspect that is how different outcomes can be explained (even without having to assume that Anthropic/OpenAI engineers are outright lying.)
But if you’ve read David Deutsch’s The Beginning of Infinity, Amodei’s view looks like a mistake. Knowledge creation is unbounded. Solving diseases/coding shouldn't result in a plateau, but rather unlock totally new, "better" problems we can't even conceive of yet.
It's the begining of Inifinity, no end in sight!
For instance, once you develop atomically precise manufacturing ala Drexler and have a complete model of biology, etc., drive solar panel efficiency to very near the upper theoretical bound for infinitely many junction cells for a raw panel of ~68%, then there isn't really anywhere to go that matters for humans. Material production would be solved, anything you could desire would be manufacturable in minutes to hours, a km^2 of solar panels could power 10-20k people's post-scarcity lives.
You eventually reach the upper bounds on compute efficiency and human upload model efficiency -- unknown but given estimates on upper bound for like rod logic (~1e-34Js/op), reasonably bounds on op speed (100MHz), and low estimates for functional uploading (1e16 flops), you get something in the zone of 0.1nW/upload, or several trillion individuals on 1m^2 of solar panel in space. When you put a simulated Banks Orbital around every star in the Milky Way in a grand sim running on a system of solar panels in space where the entire simulated galaxy has a 15ms ping to any other point in the simulated galaxy, what exactly is this infinite stream of learning? You've pushed technology to the the limits of physical law subject to the constraint of being made of atoms.
Are you envisioning that we'd eventually be doing computation using the entirety of a neutron star or (if they can exist) a quark star? Even then, you eventually hit a wall where physics constrains you from making significant further gains.
There is an ultimate end to the s-curve of technology.
But the economy didn't flatline just because we hit THAT manufacturing ceiling. Value simply migrated from manufacturing (growing wheat, assembling cars) to services (Michelin dining, DoorDash, TikTok influencers). Radio did not turn out to be the last useful invention it was predicted to be. Knowledge generation has sped up dramatically.
Your point is fair regarding hardware - eventually you do run out of stars or hit the Landauer limit. But this is exactly Deutsch’s distinction between resources (finite) and knowledge (infinite). Even in a bounded physical system, the "software" (the art, explanations, and social structures) isn't bounded by the clock speed. We don't need infinite atoms to have infinite creativity and knowledge
This is what depressed me as an early career scientist. Money to do the work to advance our species is not being distributed. Only money to generate more money for a sliver of the ownership class is distributed.
The incentives are broken. We aren’t getting Star Trek in our future. We are getting CHOAM.
In practice, quite a lot of new drugs are curative. Gene therapy, for example, usually fixes the underlying problem once and for all. Even monoclonal antibodies are rarely of the type that needs to be used for the rest of your life.
If you succeed in putting someone's cancer into remission, that patient has to be monitored for the rest of their life, but they usually don't consume any expensive drugs anymore. The expenses are more on the necessary personnel side.
There is this unpleasant fact that most chronic diseases worsen in the last 2 decades of our lives, when our systems are already seriously dysregulated by aging. Hard to fix anything reliably in a house that is already halfway down.
This is nonsense. Pharma are never in a position where they can choose between curing and treating. 90% of clinical trials fail. Pharma is throwing things at the wall and picking whatever sticks.
we're on the verge of getting to Moon and Mars in more than rare tourist numbers and with notable payloads. Add to that advancements in robotics, which will change things here on Earth as well as in space. The growth is only starting.
>The Internet also went through an exponential growth phase at the beginning.
If we consider general Internet as all the devices connected i think the exponential growth is still on as for example ARM CPUs shipments:
2002: Passed 1 billion cumulative chips shipped.
2011: Surpassed 1 billion units shipped in a single year.
2015: Running at ~12 billion units per year.
2020 (Q4): Record 6.7 billion chips shipped in one quarter (842 chips per second).
2020: Total cumulative shipments crossed 150 billion.
2024 (FY): Nearly 29 billion ARM chips shipped in 12 months.
2025: Total cumulative shipments exceeded 250 billion.I talked with my coworker today and asked which model he uses, he said Opus 4.6 but he said he doesn't use any AI stuff much anymore since he felt it makes him not learn and build the mental model which I tend to agree a bit with.
I will say that doing this for enough months has made my ability to pick up the mental model quickly and to scope how much need to absorb much quicker. It seems possible that with another year you’d become very rapid at this.
This is a key insight, I'm unable to get around this.
It's the thing I require to have before I let go, and I want to make sure it's easy to grasp again aka clear in the docs.
Basically - the sys architecture, the mental model for key things, even the project structure, you have to have a pretty good feel for.
Wait a bit longer and the next thing that's let go after you "let go" is you.
Ah yes, I feel this too! And that's much harder with someone else's code than with my own.
I unleashed Google's Jules on my toy project recently. I try to review the changes, amend the commits to get rid of the worst, and generally try to supervise the process. But still, it feels like the project is no longer mine.
Yes, Jules implemented in 10 minutes what would've taken me a week (trigonometry to determine the right focal point and length given my scene). And I guess it is the right trigonometry, because it works. But I fear going near it.
I think I'm better off developing a broad knowledge of design patterns and learning the codebases I work with in intricate, painstaking detail as opposed to trying to "go fast" with LLMs.
There is something about our biology that makes us learn better when we struggle. There are many concepts on this dynamic: generation effect, testing effect, spacing effect, desirable difficulties, productive failure...it all converges on the same phenomenon where the easier it is to learn, the worse we learn.
Take K-12 for instance. As computing technology is further and further integrated into education, cognitive performance decreases in a near-linear relationship. Gen Z is famously the first generation to perform worse in every cognitive measure than previous generations, for as long as we've been recording since the 19th century. An uncomfortable truth emerging from studies on electronics usage in schools is that it isn't just the phones driving this. It's more so the Duolingo effect of software overall emulating the sensation of learning without actually changing the brain state. Because the software that actually challenges you is not as engaging or enjoyable.
How you learn, and your ability to parse, infer, and derive meaning from large bodies of information, is increasingly a differentiator in both the personal and professional worlds. It's even more so the case when many of your peers are now learning through LLM-generated summaries averaging just 300 words, perhaps skimming outputs around 1,000 words in length for "important information". The immediate benefits are obvious, but the cost of outsourcing that cognitive work gets lost in the convenience.
Because remember, this isn't just about your ability to recall specific regex, follow a syntax convention, or how much code you ship in an hour. Your brain needs exercise, and deep learning is one of the most reliable ways to get it. Doubly true if you're not even writing your own class names.
What I am speaking to is not far away or hypothetical, either. Because as of 2023, one in four young adults in the United States is functionally illiterate.
https://www.the74million.org/article/many-young-adults-barel...
It’s also shorthand for “the end of massive R&D capex” and “the transition to market capture”. The final stage, what McKinsey types call “harvesting”, is probably not on Amodei’s radar. Based on what I’ve seen of his public personality, he would see it as too philistine and will hand it off to another custodial exec.
>To me, it is absolutely wild that you have people — within the bubble and outside the bubble — talking about the same tired, old hot-button political issues, when we are near the end of the exponential.
My interpretation is "It's pointless to discuss the old political issues, because they're not going to be relevant once AGI is achieved". So if he does believe in a plateau, it either contradicts his other prediction (that AGI will be reached in a year or two), or he believes it will plateau after AGI is already reached, which means it's kind of a pointless statement. The important thing w.r.t. all our problems being solved would the advent of AGI, not the plateau.
A large language model like GPT runs in what you’d call a forward pass. You give it tokens, it pushes them through a giant neural network once, and it predicts the next token. No weights change. Just matrix multiplications and nonlinearities. So at inference time, it does not “learn” in the training sense
we need some kind of new architecture to get to next gen wow stuff e.g differentiable memory systems. ie instead of modifying weights, the model writes to a structured memory that is itself part of the computation graph. More dynamic or modular architectures not bigger scalling and spending all our money on data centers
anybody in the ML community have an answer for this? (besides better RL and RHLF and World Models)
It learns because it remembers the context. The larger the context, the better the capabilities of the model are. I mean just give it a try and see for yourself - start building a feature, then next feature, then the next one etc. Do it in the same "workspace" or "session" and after few days, one or two weeks of writing code with the agent, you will notice that it somehow magically remembers the stuff and builds upon that context. It becomes slower too.
"Re-learning" is something different and it may not be even needed.
IMHO this is really silly: we already know that IQ is useful as a metric in the 0 to about 130 range. For any value above the delta fails to provide predictive power on real-world metrics. Just this simple fact makes the verbiage here moot. Also let's consider the wattage involved...
Quoting the Anthropic safety guy who just exited, making a bizarre and financially detrimental move: "the world is in peril" (https://www.forbes.com/sites/conormurray/2026/02/09/anthropi...)
There are people in the AI industry who are urgently warning you. Myself and my colleagues, for example: https://www.theregister.com/2026/01/11/industry_insiders_see...
Regulation will not stop this. It's time to build and deploy weapons if you want your species to survive. See earlier discussion here: https://news.ycombinator.com/item?id=46964545
(a) Top labs quietly signing deals for military deployment of frontier models in unmanned strike weapons?
(b) Top labs agreeing to license LLMs for social engineering/propaganda ops?
(c) Models that vastly exceed human intelligence and have capacity to pursue own agenda (i.e. runaway intelligence)?
(d) Something else?
It looks like dangers of AGI are overblown (perhaps partially due to grant funding and ability to get political traction/investment/competitive advantage), while (a) and (b) are severely underdiscussed. Would love to get other perspectives.
See CNN article linked here and follow links to articles mentioned in it for more details - https://news.ycombinator.com/item?id=46997198
Hinton understands the dire nature of the threat but overestimates the value of regulation in a world where the threatening technology is under development world-wide. We think regulation is basically impotent and large-scale information weapons are more viable as a solution.
This is not the time to be a docile onlooker and we urge you to take action.
https://www.julian.ac/blog/2025/09/27/failing-to-understand-...
It never does. The progress curve always looks sigmoidal.
- The beginning looks like a hockey stick, and people get excited. The assumption is that the growth party will never stop.
- You start to hit something that inherently limits the exponential growth and growth starts to be linear. It still kinda looks exponential and the people that want the party to keep growing will keep the hype up.
- Eventually you saturate something and the curve turns over. At this point it’s obvious to all but the most dedicated party-goers.
I don’t know where we are on the LLM curve, but I would guess we’re in the linear part. Which might keep going for a while. Or maybe it turns over this year. No one knows. But the party won’t go on forever; it never does.
I think Cal Newport’s piece [0] is far more realistic:
> But for now, I want to emphasize a broader point: I’m hoping 2026 will be the year we stop caring about what people believe AI might do, and instead start reacting to its real, present capabilities.
[0] Discussed here: https://news.ycombinator.com/item?id=46505735
All glory to the exponential!
This is the part I find very strange. Let's table the problems with METR [1], just noting that benchmarking AI is extremely hard and METR's methodology is not gospel just because METR's "sole purpose is to study AI capabilities". (That is not a good way to evaluate research!)
Taking whatever idealized metric you want, at some point it has to level off. That's almost trivially true: everyone should agree that unrestricted exponential growth forever is impossible, if only for the eventual heat death of the universe. That makes the question when, and not if. When do external forces dominate whatever positive feedback loops were causing the original growth? In AI, those positive feedback loops include increased funding, increased research attention and human capital, increased focus on AI-friendly hardware, and many others, including perhaps some small element of AI itself assisting the research process that could become more relevant in the future.
These positive feedback loops have happened many times, and they often do experience quite sharp level-offs as some external factor kicks in. Commercial aircraft speeds experienced a very sharp increase until they leveled off. Many companies grow very rapidly at first and then level off. Pandemics grow exponentially at first before revealing their logistic behavior. Scientific progress often follows a similar trajectory: a promising field emerges, significant increased attention brings a bevy of discoveries, and as the low-hanging fruit is picked the cost of additional breakthroughs surges and whatever fundamental limitations the approach has reveal themselves.
It's not "extremely surprising" that COVID did not infect a trillion people, even though there are some extremely sharp exponentials you can find looking at the first spread in new areas. It isn't extremely surprising that I don't book flights at Mach 3, or that Moore's Law was not an ironclad law of the universe.
Does that mean the entire field will stop making any sort of progress? Of course not. But any analysis that fundamentally boils down to taking a (deeply flawed) graph and drawing a line through it and simplifying the whole field of AI research to "line go up" is not going to give you well-founded predictions for the future.
A much more fruitful line of analysis, in my view, is to focus on the actual conditions and build a reasonable model of AI progress that includes current data while building in estimations of sigmoidal behavior. Does training scaling continue forever? Probably not, given the problems with e.g., GPT-4.5 and the limited amount of quality non-synthetic training data. It's reasonable to expect synthetic training data to work better over time, and it's also reasonable to expect the next generation of hardware to also enable an additional couple orders of magnitude. Beyond that, especially if the money runs out, it seems like scaling will hit a pretty hard wall barring exceptional progress. Is inference hardware going to get better enough that drastically increased token outputs and parallelism won't matter? Probably not, but you can definitely forecast continued hardware improvements to some degree. What might a new architectural paradigm be for AI, and would that have significant improvements over current methodology? To what degree is existing AI deployment increasing the amount of useful data for AI training? What parts of the AI improvement cycle rely on real-world tasks that might fundamentally limit progress?
That's what the discussion should be, not reposting METR for the millionth time and saying "line go up" the way people do about Bitcoin.
[1] https://www.transformernews.ai/p/against-the-metr-graph-codi...
Restated, if you let the best LLM chomp on a task for 10 hours, the output becomes slop.
* These tasks are of the type that you spend 1% of your SWE career working on.
* Each task is primed with an essay length prompt.
* You must play needle in the haystack for bugs in 10 hours worth of AI generated slop.
My experience trying AI coding at work and my observations of AI evangelists makes me believe AI coding is exclusively the purview of people who willing to handhold an AI at half pace to achieve the same result while working on software which amounts to greenfield/toy problems.
The danger of LLMs to thought work is enormously overstated and intentionally overhyped. AI : StackOverflow :: StackOverflow : graybeard in basement
It would be cool if AI kills all thought work, but what will actually happen is a undersupply of SWEs and a second golden age of SWE salaries in like 15y.
It wouldn't because of things which are not of a technical nature. Trials for instance might take much longer than developing the drug. Or complying with whatever regulations of given industry which normally requires some heavy-lifting process which is normally run by humans.
AI marketing is dystopian. They describe a world where most people are suddenly starving and homeless, and just when you start to think “hey this sounds like the conditions to create something like a French Revolution but where Bastille is a data center” they pivot to BUY MY PRODUCT SO YOU DON'T GET LEFT BEHIND.
It’s advertising straight through the amygdala.
I have no idea if they actually believe this. But it’s repulsive behavior.
The fact that Nick Land has taken hold as a philosopher in some circles in Silicon Valley truly scares me: https://www.compactmag.com/article/the-faith-of-nick-land/
>French Revolution but where Bastille is a data center”<
Even in a world where the software is 100% written by AI in 1 millisecond by a country of geniuses in a data center, humans still need to have their hands firmly on the wheel if they won’t want to risk their businesses well being. That means taking the time to understand what the AI put together. That will be the bottleneck regardless of how fast and smart AI is. Because unless the CEO wants to be held accountable for what the AI builds and deploys, humans will need to be there to take the responsibility for its output.
What happens when businesses run by AIs outperform businesses run by humans?
I doubt that we would get into a world where a company would be allowed to run without human involvement (AI directors and AI management) as you will have nobody to hold accountable.
The end of the exponential means the start of other models.
Pretraining + RL works, there is no clear evidence that it doesn't scale further.
AGI happens when you DON'T need to scale pertaining + RL.
Yet news and opinions from that world somehow seep through into my reality...
Anthropic is doing good work but he's personally responsible for a good deal of the Irrational Exuberance that plagues the space
The funny thing is his questions to her were terrible. But she rescued it anyway.
But I think he has improved markedly as an interviewer I will say
For example, at some point I grew very tired of the superficiality of the questions that Lex Friedman asks his very technical guests. He seems to be more interested into taking the conversation into a philosophy freshman's essay about technology instead of talking about technology itself.
Hearing the Dwarkesh podcast was a breath of fresh air in that regard.
The marketing effect was them catching the wave at the right time, and they're just surfing the hell out of it.
He kinda reminds me the of the Alex O Conner -same age group -very smart but inexperienced with the heavy hitters
For deep dives into AI stuff google deep mind's podcast with Hannah Fry is very good (but obviously limited to goog stuff). I also like Lex for his tech / AI podcasts. Much better interviewer IMO, Dwarkesh talks way too much, and injects his own "insights" for my taste. I'm listening to a podcast to hear what the guests have to say, not the host.
For more light-weight "news-ish" type of podcast that I listen to while walking/driving/riding the train, in no particular order: AI & I (up to date trends, relevant guests), The AI Daily Brief (formerly The AI Breakdown - this is more to keep in touch with what's released in the past month) and any other random stuff that yt pops up for me from listening to these 4 regularly.
The IPO hype is in full swing.
I personally liked that he stayed away from navel-gazing in politics when the blogosphere/podcasts went pretty heavy into that.
It did very well on twitter with a large number of high-follower-count tech people, and soon to be high-follower-count (basically AI employees). He had followed the zeitgeists general wisdom well (bat signal, work in public, you-can-just-do-things, move-to-the-arena, You-Are-the-Average-of-the-Five-People-You-Spend-the-Most-Time-With, high-horsepower). And he's just executed very well. Other people have interviewed similar people and generally gotten lower signal content. This moxie marlinspike interview is great though - https://www.youtube.com/watch?v=cPRi7mAGp7I .
And that girl Altoff ...
Literal nobodies suddenly interviewing Elon Musk, etc... within weeks.
Things rarely go "viral" on their own these days, everything is controlled, even who gets the stage, how the message is delivered, etc... as you have noticed.
With regards to who's behind, well, we might never know. However, as arcane as it might sound, gradient descent can take you close to the answer, or at least point you towards it.
I like this recent meme of Christof from Truman Show saying things like "now tell them that there's aliens" or crap like that.
Lex Fridman is a research scientist at MIT. <https://web.mit.edu/directory/?id=lexfridman&d=mit.edu>
If you want to see the mask slip, watch Lex's interview with Zelensky.
> 100% of today’s SWE tasks are done by the models.
Maybe that’s why the software is so shitty nowadays.
I do think he was overstating the current state of the models by a bit, but this is taken out of context. He is not saying this is where the models are at today.
He gives a spectrum [18:30] of the models taking over the SWE jobs:
- Model writes 90% of code (today)
- Model writes 100% of code
- Model does 90% of today's SWE tasks (end-to-end)
- Model does 100% of today's SWE tasks
- The SWE job creates new tasks that didn't exist before
- Model does the new SWE tasks as well (90% reduction in demand for SWE)
This experiment is going to fail. I only hope SWEs finally grab their balls and accept the social contract has been fundamentally broken and that they should not treat their employers so kindly next time.
The best models already produce better code than a significant fraction of human programmers, while also being orders of magnitude faster and cheaper. And the trendlines are stark. Sure, maybe AI can't replace you today. Maybe it will hit that "wall" people are always forecasting, just before it gets good enough to threaten your job. But that's a rather uncomfortable proposition to bet a career on.
Every time I read something from Dario, it seems like he is grifting normies and other midwits with his "OHHH MY GOD CLAUDE WAS KILLING TO KILL SOMEONE! MY GOD IT WANTS TO BREAK OUT!" Then they have all their Claude constitution bullshit and other nonsense to fool idiots. Yeah bro the model with static weights is truly going to take over.
He knows what he is doing, it's all marketing and they have put shit ton of money into it if you have been following the media for the last few months.
Btw, it wasn't many months ago that this guy was hawking doubling of human life span at a group of some boomer investors. Oh yeah I wonder why he decided to bring it up there? Maybe because the audience is old and desperate and that scammers play on this weaknesses.
Truly of one of the more obnoxious people in the AI space and frankly by extension Anthropic is scammy too. I rather pay Altman than give these guys a penny and that says a lot.
If you truly believe powerful AI is imminent, then it makes perfect sense to be worried about alignment failures. If a powerless 5 year old human mewls they're going to kill someone, we don't go ballistic because we know they have many years to grow up. But if a powerless 5 year old alien says they're going to kill someone, and in one year they'll be a powerful demigod, then it's quite logical to be extremely concerned about the currently harmless thoughts, because soon they could be quite harmful.
I myself don't think powerful AI is 1-2 years away, but I do take Amodei and others as genuine, and I think what they're saying does make logical sense if you believe powerful AI is imminent.
He will get more violent with his rhetoric
maybe if he can really (but really really) keep believing for 10 more years, we can have this discussion again around that time.
I trust altman more, at least hes not really pretending about who he is.
Citation needed please.
Also the same as with saying that "nuclear fussion unlimited energy is 20 years away"
Nobody. Nobody disagrees, there is zero disagreement, there is no war in Ba Sing Se.
> 100% of today’s SWE tasks are done by the models.
Thank God, maybe I can go lie in the sun then instead of having to solve everyone's problems with ancient tech that I wonder why humanity is even still using.
Oh, no? I'm still untying corporate Gordian knots?
> There is no reason why a developer at a large enterprise should not be adopting Claude Code as quickly as an individual developer or developer at a startup.
My company tried this, then quickly stopped: $$$
You may not owe AGI enthusiasts better, but you owe this community better if you're participating in it.
These posts are so tiring. The statement is an outright and blatant lie, because it's grift. The grifter wants to silence dissent by rendering it "non-existent", so that the grift can take the position of being a foregone conclusion. There is no dissent. The statement is outrageous, given the obvious amount of dissent in the comments, and the positive reaction of my fellow commenters to it. "AI built a browser from scratch." It did not. "AI built a compiler." It can't compile hello world. "AGI is coming & nobody disagrees." But the truth takes its time getting its shoes on while a lie already spread across the world.
It's doubly tiring since I (and I suspect, many of this) are having AI stuffed down our gullets by our respective management chains. Any honest evaluation of AI comes to the result that it's nowhere near capable, routinely misses the mark, and probably takes more time to verify its answer than it does to use. But I suspect many people are just skipping the verification step.
& it's disappointing to see low-quality articles like this make it, time and again, and it feels like thoughtful discussion no longer moves minds these days.
I'll try to express this without the snark going forward, though.
This captures my chief irk over these sorts of "interviews" and AI boosterism quite nicely.
Assume they're being 100% honest that they genuinely believe nobody disagrees with their statement. That leaves one of two possible outcomes:
1) They have not ingested data from beyond their narrow echo chamber that could challenge their perceptions, revealing an irresponsible, nay, negligent amount of ignorance for people in positions of authority or power
OR
2) They do not see their opponents as people.
Like, that's it. They're either ignorant or they view their opposition as subhuman. There is no gray area here, and it's why I get riled up when they're allowed to speak unchallenged at length like this. Genuinely good ideas don't need this much defense, and genuinely useful technologies don't need to be forced down throats.
this third option seems like the most reasonable option here? the way you worded this makes it seems like there are only these two options to reach your absurd conclusion
> like thats it
> There is no gray area here
re-examine your assumptions
> Like, that's it. They're either ignorant or they view their opposition as subhuman.
I'm going to go a bit off topic, but tech people often just inhale sci-fi, and I think we ought to reckon the problems with that, especially when tech people get into position of power.
Take Dune, for instance. Everyone know Vladimir Harkonnen is a bad guy, but even the good-guy Atreides seem to be spending their time fighting and assassinating, Paul's jihad kills 60 billion people, and Leto II is a totalitarian tyrant. It's all elite power-and-dominance shit, not even the protagonists are good people when you think about it. Regular people merit barely a mention, and are just fodder.
Often the people are cardboard, and it's the (fantasy) tech and the "world building" that are the focus.
It doesn't seem like it'd be good influence on someone's worldview, especially when not balanced sufficiently by other influences.
You hit the nail on their head.
They go out of their way to call you an "AI bot" if you say something that contradicts their delusional world view.
How much were devs spending to become a sticking point?
I'm asking because I thought it'd be extremely expensive when it rolled out at the company I work for, we have dashboards tracking expenses averaged per dev in each org layer, the most expensive usage is about US$ 350/month/dev, the average hovers around US$ 30-50.
It's much cheaper than I expected.
The rest of us believe that the human brain is pretty much just a meat computer that differs from lower life forms mostly quantitatively. If that's the case, then there really isn't much reason to believe we can't do exactly what nature did and just keep scaling shit up until it's smart.
'By powerful AI [he dislikes the baggage of AGI, but means the same], I have in mind an AI model—likely similar to today’s LLMs in form, though it might be based on a different architecture, might involve several interacting models, and might be trained differently—with the following properties:
In terms of pure intelligence, it is smarter than a Nobel Prize winner across most relevant fields – biology, programming, math, engineering, writing, etc. This means it can prove unsolved mathematical theorems, write extremely good novels, write difficult codebases from scratch, etc.
In addition to just being a “smart thing you talk to”, it has all the “interfaces” available to a human working virtually, including text, audio, video, mouse and keyboard control, and internet access. It can engage in any actions, communications, or remote operations enabled by this interface, including taking actions on the internet, taking or giving directions to humans, ordering materials, directing experiments, watching videos, making videos, and so on. It does all of these tasks with, again, a skill exceeding that of the most capable humans in the world.
It does not just passively answer questions; instead, it can be given tasks that take hours, days, or weeks to complete, and then goes off and does those tasks autonomously, in the way a smart employee would, asking for clarification as necessary.
It does not have a physical embodiment (other than living on a computer screen), but it can control existing physical tools, robots, or laboratory equipment through a computer; in theory it could even design robots or equipment for itself to use.
The resources used to train the model can be repurposed to run millions of instances of it (this matches projected cluster sizes by ~2027), and the model can absorb information and generate actions at roughly 10x-100x human speed. It may however be limited by the response time of the physical world or of software it interacts with.
Each of these million copies can act independently on unrelated tasks, or if needed can all work together in the same way humans would collaborate, perhaps with different subpopulations fine-tuned to be especially good at particular tasks.
We could summarize this as a “country of geniuses in a datacenter”.'
It's a constantly shifting goalpost. Really it's a just a big lie that says AI will do whatever you can imagine it would.
Meanwhile, Claude Code is implemented using a React-like framework and has 6000 open issues, many of which are utterly trivial to fix.
Can I ask what happened with your Claude Code rollout?
Oh good, hopefully it'll model itself after an exponential rise in any sort of animal populations and collapse on itself because it can no longer be sustained! Isn't that how things go in exponential systems with resource constraints? We can only hope that will be the best outcome. That would be wonderful.