upvote
Once they "reach AGI", will they have a big party on a carrier with a "Mission Accomplished" banner?
reply
They don't need to reach AGI. They just need to put all of the engineers on HN out of work.

A year ago I would have said that was crazy. In the last month, I've been using Claude Code to write 20kloc of Rust code every day (and I review all of it).

A week is now a day. If that figure doubles, I have no idea what will happen to us. And I think it's coming.

reply
A classic hype merchant sales pitch: believe me, I was a doubter just like you, but I saw the light thanks to [insert latest model]!

(Which for anyone familiar with your long comment history as a regular HN poster, is comically absurd to imply. You've been reliably adamant that AI will demolish this or that entire industry overnight for years at this point).

reply
> write 20kloc of Rust code every day (and I review all of it)

Only one of this can be true. It's not a shame to say you don't bother reviewing it, in the future that may well be the norm.

reply
So you now have 400Kloc of Rust code? Doing what? How much of that is "new"?

I can't get Augment / Opus 4.5 to edit a few C++ files from within VSCode without going off on a wild goose chase or getting stuck in an infinite loop after I tell that it should be doing this: "oh, you're right, I need to do X", "To do X, I must understand how to do Y", "I see now that to do Y, I should look at at Z". "Let me look at Z", followed by: "oh, you're right, I need to do X"..

reply
To be fair, humans editing C++ also go on wild goose chases. Have you seen the insanity the C++ committee has ratified?
reply
To do what, exactly, and are people paying you for your output or are you just making things for yourself?

Building things at a mature company with a market is a lot different than hacking together your own tools. There are a lot more people you can let down at scale.

reply
> 20kloc of Rust code every day (and I review all of it).

Reviewing 1k lines of code an hour is a breakneck pace, are you spending 20 hours a day reviewing code?

reply
It’s clearly code so flawless you can tell at a glance that it’s correct.
reply
What does all this code do? What software are you writing?
reply
"All of it"
reply
Best comment of the day. In three words no less...
reply
> They just need to put all of the engineers on HN out of work.

I think you've crossed the line from being an AI maxi to just rage baiting. This comment is a pointless anecdote at best, please take your ridiculous FOMO takes elsewhere.

reply
That’s the same definition of reviewing code as saying watching the movie is the same as reading the book it’s based on. No human has ever reviewed 600k lines of code in a month, ever. It’s hard to find someone who can even read and understand that amount in that time.
reply
I’m convinced these “guys you gotta believe me I’m a seasoned veteran and this shit is the real deal” posts that show up in every AI thread are either coming from Sam Altman or a bot.
reply
Just try it.
reply
Yeah, I do use agentic coding a good bit. It’s impressive, but not so much that I’ve convinced myself that my days are numbered.
reply
I have. It's great. Not 20 kloc/day great, though, and nobody believes that you are giving >1 kloc/hour anything more than the most casual glance.
reply
I had to hack my scrollbar speed to 500% to get through all the code reviews in a day
reply
It'd be interested in seeing how exactly the lawyers figured out how to define AGI. It must be a fairly mundane set of KPIs that they just arbitrarily call AGI, the term will probably devalue significantly in the coming years.

The actual quote is this though:

> hitting an AGI milestone or pursuing an IPO

So it seems softer than actually achieving AGI or finalising an IPO.

reply
I'd assume the real trigger here is "reaching AGI," which would help OpenAI shrug off some of their Microsoft commitments thus making OpenAI models available on Amazon Bedrock. Which is what Amazon is really after.
reply
Has OpenAI laid out the specific definition of what an AGI is for this case? The one from their mission is quite vague and the general community has nothing close to a universal common definition... which means they will most likely just define it as what they already have when the timing is right.
reply
At least in their Microsoft contract it means $100 billion in profit, though they don't need to have actually made that money, they just need to show they're on track to do so.
reply
Very convenient to put "AGI" in all these agreements because the term is fundamentally undefinable. So throw out whatever numbers you want and fight about it and backtrack later.
reply
> fundamentally undefinable

Incredible, how an entire religion has sprung up around AGI.

reply
The definition used to be "passes the Turing test" .. until LLMs passed it.
reply
Extremely debatable. Especially because there is no "The Turing Test" [0] only a game and a few instances were described by Turing. I recommend reading the original paper before making bold claims about it. The bar for the interrogator has certainly be raised, but considering:

- the prevalence "How many |r|'s are in the word 'strawberry'?" esque questions that cause(d) LLMs to stumble

- context window issues

It would be naive to claim that there does not exist, or even that it would be difficult to construct/train, an interrogator that could reliably distinguish between an LLM and human chat instance.

[0]: https://archive.computerhistory.org/projects/chess/related_m...

reply
Sure, when the expected monetary value was 0. Then they started claiming that investing $1,000,000,000,000.00 (that's $1T) into a 4 year old startup was a good idea. Change the valuation, change the goal. Then the goal was be better than a human employees (or at least more efficient or even just improves efficiency) because without that the value of the LLM is far lower than what it is being sold as. All the research so far says that LLMs fall far short of that goal. And if this was someone else's money, fine. But this is basically everyone's retirement savings. Again, higher valuation, higher goal. Finally, when you start losing people's retirement savings, criminal penalties start getting attached to things.
reply
The problem with AGI is not that it's undefinable, but that everyone has a different one. Kinda like consciousness in that regard.

Fortunately, OpenAI already wrote theirs down. Well, Microsoft[0] says they did, anyway. Some people claimed it was a secret only a few years ago, and since then LLMs have made it so much harder to tell the difference between leaks and hallucinated news saying this, but I can say there's at least a claim of a leak[1].

[0] https://blogs.microsoft.com/blog/2026/02/27/microsoft-and-op...

[1] It talks about it, but links to a paywalled site, so I still don't know what it is: https://techcrunch.com/2024/12/26/microsoft-and-openai-have-...

reply
All the major investments in these big rounds have come in tranches, right?
reply
So they’re getting in on the IPO.

Are they going to get stock for it or is it a PIPE?

Personally, I don’t think I want to get in on this at retail prices.

It can both be true at the same time that AI going to disrupt our world and that being an AI lab is a terrible business.

reply
like the other comment, openai can force itself onto the massive index like VOO/FXAIX etc to make retail folks to provide liquidity exit for openai investors.
reply
But will you have a choice once they enter the indexes ? People are automatically going to invest into that (circular) pyramid scheme.
reply