Whats interesting to me as well as much as companies are pushing AI adoption, i have started to hear AI token spend limits enforced across a few companies, so its not entirely clear that b2b can make them profitable yet either.
If all the models reach good enough, then low cost provider would win. Gemini seems like a safer bet since Google controls more of the stack / has more efficiencies / cross selling / etc.
It’s not like “best” has won any other b2b arms race in the past.
Gemini is the best deal too. For $20: you get multiple quotas per day across the products (web, CLI, antigravity, AI Studio) 2tb of cloud storage, and you can family share the plan.
Further they have their own TPUs, datacenters, etc on which to run their models.
Plus existing data they've squirreled away over the preceding 30 years from books, web, etc.
Just seems like a lot of efficiencies if its going to come down to cost.
And in that reality one can’t just magically spend a bunch more on some fancy new thing, especially when said fancy new thing isn’t retuning value. So “token limits” and cost controls on B2B is entirely expected here.
I think this is the key element. Either they can't measure the value, or it's far far lower than anyone wants to believe, or both.
I think the problem is less that it makes some coding tasks XX% faster, but that the end to end of a SWEs roles tasks is only improved by some much smaller Y%.
If a CTO sets $10k/year spend limits on $500k SWEs.. they must not believe any of the hype.
Expert systems were amazing. They were not cost effective.
There might be another bitter lesson to be had here, and unless the accountants start talking we're not gonna know any time soon.
Fuller integration into the user's life will bring ever more ad opportunities (and it doesn't matter if the HN base hates that notion, it's going to happen regardless). That'll happen over the next decade gradually.
Shopping, home management, tasks (taxes, accounting, lifestyle, reminders, homework, work work, 800 other things), travel (obvious), advice & general conversation (already there), search (being consumed now), gaming (next 3-5 years to start), full at-work integration (gradual spread across all industries, with more narrow expertise), digital world building (10-15+ years out for mass user adoption). And on the list goes. It's pretty much anything the user can or does touch in life.
We already have the tech for that, why hasn't it happened? People are revolted by the AI results in Google. AI isn't going to make people use their computers more. It's not opening up a new consumer market. This is just making each search infinitely more expensive.
The latest "Thinking" version gets it reliably right but spent about 3 minutes coming up with the answer that 10 seconds of googling answers.
So I don't believe we are currently in a situation where LLMs are an effective replacement for search engines.
And what do you think this'll do for future LLM models that need to train on new content if web page traffic collapses?
I think Google has several ai products with search features?
Which one in your experience "seems correct"?
I'm fascinated because I've never found any LLM to be particularly error free at search.