Also, note that there's zero CUDA dependency. It runs entirely on Huawei chips. In other words, Chinese ecosystem has delivered a complete AI stack. Like it or not, that's a big news. But what's there not to like when monopolies break down?
The US-China contest aside - it is in the application layer llms will show their value. There the field, with llm commoditization and no clear monopolies, is wide open.
There was a point in time where it looked like llms would the domain of a single well guarded monopoly - that would have been a very dark world. Luckily we are not there now and there is plenty of grounds for optimism.
When China does good, it's always that they do mostly bad.
With China it's always pointed out how much power the state has over corporations there, but in the US out of control lobying is supposed to be 'concerned citizens expressing their opinions' or some shit. We're still supposed to take for granted that it is a representative democracy, if a flawed one.
It's a small difference, but important. Especially because that person is far more likely to be responsible (voting) for and profiting from USAs bad stuff.
The issue is propagandists are typically brainwashed already.
I'm an American and I don't believe that.
With China, you can say 'yeah, this is good, but they eat babies for fun' and it would mostly pass with people nodding along.
The U.S politics are easier to understand from the outside. For one it's a democracy, a more transparent process despite a lot is happening behind curtains. I have no idea what North Koreans are able to make of the U.S scene, I know for sure people in U.S and Europe are hardly able to comment on N.K.
tldr: I'm with you non Americans (and Americans) are perfectly able to critique the U.S with some valuable accuracy.
It seems to me that there is a fair amount of misinformation which gets spread about the US. For example, many non-Americans seem to believe that school shootings are a significant cause of death here.
Furthermore, your proposed scheme creates an incentive to be non-transparent and thus not vulnerable to critique. By closing off information about your country, you can say to any critic: "Your critique is incorrect, because you lack information." Thus creating a reputational advantage for countries which successfully clamp down on the flow of information.
Is that your desired outcome? You want a world where criticizing the US can no longer be done as soon as Trump kicks out all of the foreign journalists and stops the information flow?
I.e. he doesn't see the US as "the good guys" either.
Pointing out the war threat from China isn't hypocritical just because you don't list all the war threats from the US at the same time.
Of course not. When it comes to SOTA LLMs you have the choice between two bad options. For many, choosing the Chinese option is just choosing the lesser of two evils (and it's much cheaper).
Mistral is right here, their models are in-between the cheap to run Chinese models and top of the line performances of US frontier models.
The safe money is they are going to be an also-ran for the AI revolution. They did manage to force Apple to switch from using lightening connectors to USB though so their wins can't just be laughed off. Maybe they'll surprise us but it'd be a welcome change from their usual routine.
Production of state of the art semiconductors, yes. NXP, STMicro, Infineon are still there and massive in automotive, industrial, card chips, etc.
> The EU fumbled the software revolution, the successes mainly came from the US
Worldwide massive success, mostly yes. Most European countries have their local or regional success stories though.
> The safe money is they are going to be an also-ran for the AI revolution
Not really. Past performances, or lack thereof, are not indicative of future ones.
Mistral are pretty good and selling well in the enterprise space. Some of the best voice models are coming from France (Kyutai).
If you fall out of the state of the art then the claim of EU fumbling semiconductors is correct.
>NXP, STMicro, Infineon are still there and massive in automotive, industrial, card chips, etc.
The EU semi companies you listed are absent from the state of the art and only make low margin commodity parts that don't have moats. No, ASML is not enough for claiming EU superiority since the EUV light source is still US IP designed and manufactured.
>Worldwide massive success, mostly yes.
Worldwide success is where the big money is, and you need a lot of money for cutting edge research and experimentation to build the future successes. Hence the claim of EU fumbling software is correct.
>Most European countries have their local or regional success stories though.
EU mom and pop shops aren't gonna make enough money to be able to afford risky ambitious ventures the likes of FAANGs have. Which is why you work for Hashicorp, a large global US company, and not some local EU company.
Then they need money.
So most of the talent flee or get bought, typical example in machine learning space is huggingface or fchollet.
Then European government plays catch-up and offer subventions, but at the same time makes rules to make sure companies don't threaten US dominance, or Asian manufacturing.
Mistral is typically playing catch the subsidy game.
Europe is constructed so that it can't win, but can "pick" the winner between scylla and charybdis, pest and cholera.
Because they have no spine and no leverage/muscle on the international stage to throw their weight around and make sure they get what's best for themselves at the expense of everyone else the same way US, China, etc do.
They play the international nice guy that just ends up being the doormat everyone takes advantage of, being at the mercy of Russian and Azeri gas, at the mercy of US tech, energy and defence, and at the mercy of Chinese manufacturing after dismantling their own manufacturing, at the mercy of Turkey for migration enforcement, etc so they can't do anything radical that upsets either or risk massive repercussions they aren't prepared for, so they just turtle, bury their head in the sand and pretend everything is going fine.
There's nothing virtuous in weakness.
Mistral is good for many tasks where you do not need SOTA or near SOTA performance. They cannot compete if you do.
When someone points out hypocrisy, this is "the answer", it seems. But it is just a statement, not a rebuttal of the hypocrisy that was pointed out.
Hypocrisy is still hypocrisy.
And bad things are bad things. Yet no amount of propaganda (red scare, "eew dictatorship", Uyger-genocide, Taiwan threat) can convince me that the China is as evil (or more evil) than the US-Israel alliance of the the last 50 years.
Not mentioning US problems every time they criticize CCP problems is not automatically hypocrisy, and this idea basically means you cannot criticize anything without criticizing everything someone considers just as bad or worse at the same time.
Calling a discussion on China hypocritical because it doesn't say "but US worse" is essentially trying to build in whataboutism into every discussion.
It's a symptom of increasing polarization and part of the problem.
https://www.youtube.com/watch?v=P7W20hdgWXY
I think I'll take the open AI models, innovative high quality EVs and cheap solar panels, please.
We conduct amoral behavior with terrorist regimes for dollars.
The current president - who Americans voted for twice - is heavily accused of being a pedophile and has reneged on every one of his poll promise
Really not the best advertisement for democracy
Why would Russians want democracy? Or the Chinese, for that matter? There have been zero democratic impulses in their societies across hundreds, even thousands of years.
The west needs to rest its democratizing mission and accept that every society is fundamentally different
My country (India) got a "thriving" democracy, but because there is no real democratic impulse in the society, everything on the ground has devolved into what the society was always like - quasi-feudal bureaucracy
Well, ideology. I believe my way is the only way for every population in the world too, and I fight for it to happen. Of course, each place adapts to their own condition, but I believe my core ideology is the way for humanity as a whole, and I believe it is the same for people who defend western american-style democracy.
The marched for it en masse in 1989?
Russians and Chinese are also people. They deserve to rule themselves.
They are ruling themselves in the sense that their governing systems are emergent consequences of their own cultures. All peoples ultimately deserve the governments they have.
Guess the Tiananmen square tank man is a victim, but Alex Pretti and Renee Good are just statistics
(The tank man wasn't even run down by the tank - Good was shot for merely turning the wheels in the wrong direction)
Americans really need to shut up about any democratic values or humans rights and clean up their own mess before preaching to the world
It's not Americans, it's educated people who believe in personal liberties.
> Why would Russians want democracy
Because they would have a choice if they want to be robbed blind by a bunch of oligarchs, and if they want to be sanctioned off from the world because the supreme leader decided he wants to kill and maim a million Russians to achieve nothing more than killing Ukrainian civillians.
> There have been zero democratic impulses in their societies across hundreds, even thousands of years
Absurdly bad historic revisionism. Russia had democratic impulses in 1917 and 1990, both hijacked and went nowhere. China's 1911 revolution was also overtly democratic in nature, but was also hijacked.
I find this attitude deeply parochial and colonial. Who are these so-called "educated people" (most of whom would be in western developed nations) to decide what sort of governance system a country should have?
The democratic revolution in America and France came from its own people. If the Russians or the Chinese want democracy, they'll get it on their own
Western hand-wringing about the "lack of democracy" in foreign (usually poorer) countries is just concern-colonialism. I think most of these educated people should focus on their own countries and let the rest of the world be
I think a much better metric is suppression of dissent, human rights records etc., not (the illusion of) choice at the poll booth once every 4 years.
Also, consumer goods.
The voting and multiple-branches-checks-and-balances elements are sidelines.
Currently none of those promises are true in the US. The government is murdering and jailing people for whimsical and self-indulgent reasons, the consumer economy is about to crash, and the only checks-and-balances are the checks going straight to the Emperor's private accounts.
To be fair, there's some judicial pushback, and some political friction.
But Senate and Congress are wholly captured, the opposition is flaccid and foreign-funded, media independence is a myth, and the last time The People had any real influence on policy was the 70s. Possibly.
I have no idea if China is "better". From a distance China seems to be doing much better at building useful things and making long term plans.
But ruling cliques always seem to end up being run by psychopaths, so my expectations for humanity from China's rulers aren't any higher than those for the US.
It would be hilarious if it wasn't so sad
It turns out that the people will vote for some terrible things in order to get that one petty little thing a given candidate promises and they want, or because they don't like something specific about the other candidate(s). And of course many may later say “well, I didn't vote for that” when they quite demonstrably did.
The problem is that people put stock in pre-election promises, rather than voting for the character of the person they want to represent them.
The measure is the number of votes. "What shall we have for dinner" measures things, there's no target in a "curry vs pizza vs thai" poll, and it doesn't really matter, the target is a nice night in with a film.
However with politics, getting power is the goal, thus the number of votes is thus the target, and thus its not good at measuring what the country actually wants, just who can best get the most votes.
This isn't new, but modern brainwashing allows manipulation at a scale hitherto unseen.
The reality is that the term democracy in western society has essentially become meaningless due to the swathes of algorithmic manipulation which occurs every second of everyday through every possible digital medium.
Western democracies don't have that problem. Yes, they have other problems. Many problems which are hard to solve. But if you live in a western democracy you can freely criticise those in power without fear of retribution.
Right.
The name says "demos" and "kratos" but names are names, not facts.
There are many ways to give people a choice and this one has proven to be quite ineffective at that, as it slowly devolved into a plutocracy/oligarchy. Iron law of oligarchy, yadda yadda.
What they are very effective at though: crushing dissent, calming the masses with a reassuring illusion of choice, and touting itself as the "one true way".
When I look at the outcomes I don't see any semblance of democracy, only a ritual dance/theatre show every 4 years. A farce as big as the "democratic" instruments on the PRC.
There's a reason this "democracy" is very diligent at discouraging association and unionizing. Those give actual power to the people (and with power comes choice). That's dangerous. People might start believing they can actually influence the outcomes.
"Don't blame me - I voted for Kodos"
Do not conflate the broken American political system, the semi-broken British one, and the whole rest of the "west". Each country has its own political system, and they are wildly different.
> crushing dissent
Democracies are good at crushing dissent? Compared to other political systems? That's just not true. All other political systems rely on universal truth and unwavering trust in a person / religion / clique of people, who can do no wrong and can never be criticised.
> There's a reason this "democracy" is very diligent at discouraging association and unionizing
What? You are probably talking about a specific democracy, and the most broken one at that.
As someone from the "whole rest of the west", no, they're not different at all. Very minor details change, but the net outcome is the exact same and suffer from the exact same problems.
You can't escape the iron law of oligarchy.
> Democracies are good at crushing dissent?
They're not only good: they are the best. You don't need to curb dissent by violence if you discourage dissent by social manipulation. It's the cheapest and most effective tactic: keeping the populace docile.
If you manage to equate "democracy" (again, quotes intended) with democracy (lack of quotes intended), most of the work is already done.
"What are you, antidemocratic!?"
"Don't blame me - I voted for Kodos"
There's a reason my country's system trembled when the bipartisan system was challenged as new parties emerged... but it was curbed within two legislatures without a single shot fired and now we're back to an even stronger bipartisan representation. Quite the fine job, actually.
Assuming that everyone who disagrees with you is a propagandized bot is a terrible way to live. You will not learn.
Neither is the US, land of slaves, segregation, and the KKK. They did seem to get better there for a few of decades, but sure are working hard to return to their roots.
Isn't the US building mass detention camps right now for all the brown people there? And arresting / detaining / demanding papers from any and everyone? With federal agents killing civilians?
Don't get me wrong, China is also horrible here, they have their own camps.
But pretending the US is positive wrt human rights is a wild take in 2026.
No, it is not, but the freedom of speech protections the US has (that China doesn't) allow for such commentary.
And yes, they are-
https://en.wikipedia.org/wiki/List_of_immigrant_detention_si...
Why would you think that?
> And arresting / detaining / demanding papers from any and everyone?
I have lots of friends from outside the U.S. that come regularly and don't find it onerous. Maybe it depends where you are coming from?
> With federal agents killing civilians?
OK, I agree that there are issues, and even very serious ones. Obviously, not on the level of China, but still serious issues. Nonetheless, what you see on left leaning media is not representative of what is happening on the ground throughout the U.S. Not even close.
IMO, the US is definitely positive wrt human rights. There are issues, but you can go to a No Kings protest, and live your life happily without issues, and it is hard to find another country that is nearly as forgiving. And it at least has people trying to spread concepts of individual liberty, vs most countries in Europe, almost all countries in Asia, and ALL Muslim countries, that are leaning to removing individual rights.
No? Its for illegal people, regardless of color. Just so happened that most illegals come from specific places
If you meant American citizen human rights, then you’re correct.
Not even that. ICE has already killed US citizens, they no longer prohibit segregation, trans people were banned from the military, and many more. All of those affect American citizens.
How about your pack up your arrogance and stop defining human rights for me and other 1.4 billion Chinese?
It's not like 1.4 billion Chinese have much say in that.
If I am wrong, please remind me again how much say Chinese people had on the escape hatches of Article 51 in your constitution.
Ask around in Vietnam, Iraq, Syria and countless more countries around the world.
They didn't even say that. They only said China playing is "better than leaving everything to the US alone."
The US was one of the first democracies in the world, and many countries followed suit. But the US hasn't kept up, and now the powers that be have exploited the weaknesses in the system. With arguably the biggest one being giving the president too much power (appointing supreme court justices, executive orders, etc).
Don't get me wrong, I'd rather live in a country without a million cameras that automatically fine me for crossing the street illegally but I don't actually deceive myself in thinking my vote counts for much.
Are you talking about the US or China? https://deflock.org/
China at least banned the use of facial recognition in public spaces by their supreme court in 2021 (and then further strengthened the ban in 2024 and also got the PIPL).
If you're thinking of the "social credit" system please know that that's just an online meme. China's credit score system is not even nationalized and not nearly as invasive as the US's credit score system, which can sometimes determine whether or not someone is allowed to buy a house.
Besides their own credit score system, the other thing that sometimes gets labelled the "social credit system" was an attempt they had to track the behavior of business leaders and elected politicians. Basically anyone who holds social power but not the common person. This also never really took off and was not ever nationalized/centralized.
Agreed, but there again, the democracies have surveillance capitalism, it's not exactly like we're not being tracked.
The elected government of the US has the moral highground of over the regime that killed the KMT in it's weakened state after the KMT defeated Japan, went on a rampage against the educated classes, mowed down its own people with machineguns and tanks when they demanded a say in their own governments, and kidnaps people advocating for democracy to this day, including Jack Ma.
> despite starting a new war... on behalf of Israel every six months.
The war started when Hamas, funded by Iran, went on a murder and rape rampage against Israeli civilians.
I am not washing away the authoritarianism, but take a look at other economic super powers directionality. Or that of tech ceos as well. At least Chinese tech companies aren't going around praising wwii Germany, writing manifestos, and bombing children at school or fisherman on whims. It is difficult not to see more countries regardless of leadership putting their hat in the ring as a net positive. Especially if it increases sustainability and lowers the price, which this very clearly does. It's even open source...
Fully agree. From a US perspective, that sucks. For everyone else it's pretty great.
At this point the world's opinions of China are better than those of the US in some polls. One country invests and helps build infrastructure on a massive scale globally, the other alienates allies, causes countless conflicts, and openly threatens to end civilizations.
Indeed, even if one isn't partial to China, there's reasons to be glad that an increasingly hostile US has powerful competition.
> This is about who will dominate the world of tomorrow.
For this you'd need a technological moat. So far the forerunners have burned a lot of money with no moat in sight. Right now Europe is happy just contributing on research and doing the bare-minimum to maintain the know-how. Building a frontier model would be lobbing money into the incinerator for something that will be outdated tomorrow. European investors are too careful for that - and in this case seem to be right.
My point is that Trump could sign/execute/order all the same exact things he's done, but if I just never spoke about it, or kept hidden like Chinese do, he would be compared MUCH differently.
That would also make him a lot more dangerous. After all in his first presidency he was still the man behind the biggest military on the planet but he knew shit on how to leverage this. In his second term he is even more loose but loose is tempertantrums and simple short sighted strategies. Easy to read, hard to accept.
In the US its not the Uighurs or Tibetans who are being oppressed - it's the blacks and immigrants. The US elected a president who characterizes immigrants as rapist and murderers (while he himself is a convicted rapist, suspected pedophile, and wants to commit war crimes in Iran).
The facade, believed by many Americans, is that USA is the land of the free, a democracy (despite no popular vote) one of the good guys, but actions say otherwise.
China's policies and government aren't morally defensible and I do fear that they will become more aggressive in spreading their influence and policies onto other countries, but from an economic standpoint what they're doing is super effective. While the previous world power (the US) is stuck in infighting and going through cycles of fixing/undoing the previous administration's damages, instead of planning ahead.
It’s this sort of example (and not properly supporting Ukraine, and not agreeing how to collectively deal with migrants, and not agreeing how to coordinate defence, and myriad other examples) that highlights what a pointless mess the EU is. It’s not a unified block - it’s 27 self-interested entities squabbling and playing petty power games, while totally failing to plan for the future with vision.
The EU could/should have ensured that a European equivalent to OpenAI or Anthropic could thrive, and had competitive frontier models already; instead, they’re years and countless billions behind.
Which is crazy given that ASML is European.
I don't know what the problem is. Are we europeans to stupid? Do we just not have enough money / VC money? Are we not proud enough?
:(
Shared language and history aside, these two cultures are not in the same solar system when it comes to social norms and curtesies.
That just goes to prove that Taiwan is the real China and the others are just communist barbarians.
I think they are leaders in the democratization of LLMs. Almost everyone has a computer right now that can run a useful variant of a Mistral model. I hope they keep their focus because what they are aiming for likely has the biggest impact on the average person and would be the best case scenario for the technology in general.
Their main selling point is: They are neither US-American nor Chinese. That's a real moat in today's world. I think at the moment they feel quite comfortable.
They sanctioned the hell out of Huawei and now Huawei is bigger than ever
America is just not able to digest the idea that another country can be as good, if not better, at innovation
China's fall in the 19th century came at them for the same reason. How could these European savages be stronger, thus better than us? Our intelligence service must be out of their mind.
Walmart is a horrible company owned by horrible people and yet it’s cheap so it dominates.
If the quality really is in the Opus 4.6 range (considering how bad 4.7 is), then it’s a pretty big deal.
its naive to think they would have stayed on a 'western' stack.
Most of the time 'losing' isn't making a bad choice its being put in a situation where you have no good choices.
China is not perfect but a bit of competition is healthy and needed
The next decade is going to look very different with America Alone.
With all that goes on it has changed. Recently I sat on a plane near some Americans discussing their holidays here, and I noticed I felt contempt. Sitting their with insane privilege as their government torches the world.
Individuals remain individuals, and one really ought not to be prejudice. However the lack of resistance I see in in the “land of the free” as their “democratic” institutions collapse just makes me believe they never cared at all. In France cars are torched if the pension age is raised. In America the rise facism apparently doesnt matter to them.
Largest protests in US history just in the past year:
https://en.wikipedia.org/wiki/List_of_protests_and_demonstra...
>insane privilege
My sister and brother recently graduated from college, have been searching for jobs for over 6 months, they can't find anything. They're politically liberal Californians.
My family in law seems to swing slightly republican. As a Dutchie, I could get some answers because I'm too naive not to talk about politics. So I got to probe a bit. What I simply found was that they'd say "I can't trust the news, none of it. Not CNN, not Fox News, nothing". Then I'd say "well in the Netherlands, I'd argue that while news outlets have their bias, you can trust them on basic factual reporting". She looked at me with a stare that I could only describe as "oh but honey, you're too young and naive to understand". To which I thought "you don't know the Netherlands. We're not perfect but we're nowhere near as deranged as what I'm seeing here".
I think that explains a lot of it for some people. The trust in the media, all media, is completely broken. Trump has how many fellonies now? Can't trust it. Kamala is doing what now? All talk. DOGE is fixing the government? I fucking hope so! But can't trust the damn news. Whether they do or don't, they are always burning money, god damn bureaucrats.
I feel that's the mindset that my family in law has.
This view gets echoed here on HN a lot. I find it very strange to be honest, because I tune in to CNN and I see lots of bias in the commentary and editorial, but when it comes to factual reporting they are pretty straightforward and down to earth. It seems to me that the real issue is people don't seem to distinguish between reporting and editorial content / commentary. Stop watching that garbage and actually consume the factual content and analysis. Yeah it's dry and boring but if that isn't enough for you then it just shows you never cared about facts in the first place.
No, not really. I mean for me, yea, sure, easy. But in the general case? It depends on who you are.
The reason I trust CNN is because when a Dutch news source reports more or less the same thing, I can easily see the reporting matches with that of CNN. Because of this, I personally have some built up trust with CNN. When I look at Fox News, oh deary... it's nothing like what I see on the Dutch news.
This is not something I do consciously, it's simply that I happen to watch Dutch news sometimes and I happen to see American news sometimes and it costs no effort for me to compare. Combine that then with that on HN I also sometimes see BBC and similar British venues (e.g. The Economist is also British I believe?), and now I suddenly have 3 countries worth of news sources.
Many Americans don't really know that the UK exists other than that they rebelled against it. Many Americans almost haven't left their 20 mile radius world (many also did of course). But it's these people that I tend to have a lot of in my in-law family or however you call it (schoonfamilie in Dutch). I'm quite exotic to them in that sense, and definitely foreign. Thank god they have some Dutch roots.
Point being: with that mindset, you're not checking out what the BBC has to say on a topic. You're checking American news, not because of patriotism but simply because of that's all you know and going outside of what you know costs effort. And you already have a job to do, come home late, just want to watch your shows in the evening and that's it.
I am by no means saying that this is representative for all Americans, it isn't. What I am saying is: I see this a lot in my slice of the US. The reason I'm sharing it is because what my in-law family is saying is definitely at a much more personal level than whatever conversation I've had with some random, but lovely, person from a hacker space or hacker house in San Francisco.
Yet, I don't see this view a lot on the news. Nor do I hear Dutchies talking about it, they are simply out of the loop when it comes to a view like this. I don't know how prevalent it is, but if many people of a family of 50 to 100 people is in a situation like this, then my bet is that they aren't the only family.
My running hypothesis has been the trust breakdown arises from social-media overexposure driving lazy nihilism, which in turn gave free reign to a uniquely-corrupt class of politicians. But I'm not sure how to neutrally evaluate that.
The most famous examples are likely the tobacco industry spreading misinformation through self-funded studies and experts, and the fossil fuel industry doing the same to seed doubt about climate change. But of course we can think of countless examples of entire industries and individual large corporations pushing out misleading bullshit, threatening or outright killing journalists and activists to cover up their catastrophic fuckups and their chronic conscious excretion of negative externalities.
This has all of course been going on since the dawn of time, but to focus on the last century in the US, we've seen all sorts of corporations and coalitions of rich and powerful people push misinformation into nearly every sector of our society - universities, science, journalism, politics, etc. in order to undermine confidence in shared facts, corrupt people's ability to discern whether or not something is fundamentally true, and sow confusion so that they can continue to operate in perpetuity in this chaotic maelstrom of doubt.
Lots of capture of government towards these ends as well, we can look at the concomitant constant cuts to education in order to weaken people's understanding of the world and ability to think critically. The revocation of the Fairness Doctrine was probably a step change, and Trump represents the sharpest recent escalation of all this.
From day one, he's done everything he can to shred any collective notion of shared objective truth. Anything he doesn't like is fake news, and the idea that the media is lying, scientists are lying, experts are lying, and institutions are lying, he has spread so fucking successfully through society, to the point where Americans no longer have anything like a shared sense of reality.
It seems like we're being reduced to tribes who are organized primarily around faith in various charismatic individuals.
I think this is fundamentally the worst thing he's done, because it lays the foundation for virtually every other conceivable and inconceivable abuse. If people can't even agree on what is happening, we're fucked. People and institutions in power can do anything they want to whoever they want, because the public has lost their ability to even recognize the danger posed to them collectively and thus mount any resistance based on a shared sense of reality.
Social media has definitely famously accelerated aspects of this like the fragmentation and the spread/magnification of fringe worldviews through echo chambers, but I think it's just one (and maybe this is controversial, but I'd be willing to be generous enough to think the 20something year old creators were too stupid to conceive of these long term consequences at first, but who knows, maybe not) element in a much longer and more intentional, malicious war against the many for the benefit of the few.
Shooting from the hip here. Feels like a duct tape hack on first thought.
I mean that's what I do, subconsciously. I think a lot of Europeans do this because a lot of Europeans tend to speak English and then their actual native language, or something similar (e.g. I wonder how Swiss people experience this).
As a European, how do you influence your government?
And you’re right, most Americans do not understand the privileges they have or give one single shit about democracy; it is just not a salient political issue. But eggs… don’t get me started on eggs.
I feel like the issue there is that alarm bells in of themselves solve nothing. I won't extend that argument to one of its obvious conclusions, but instead I will say that efforts to attack education and critical thinking skills all contribute to people being susceptible to their democracy being corrupted and robbed blind - so having an educated populace with a sense of integrity and respect of human rights would help!
This is why the swing voters / swing states are so important in the US, because only a few million are flexible enough to switch sides.
Of course the core issue is that there's a two party system; while I'm sure that in a healthy democracy the current republican and democrat parties would be the bigger ones, they wouldn't have a majority.
Of course if the USA was an actual democracy, electing it's president by popular vote, then this would not be an issue - every vote would count to tip the balance in favor of who the people wanted to elect, not just the votes of the 20% fortunate enough to live in a "swing" state.
This, for me, is the crux. Politics is treated like a team sport in the US, you pick your side and cheer them on no matter what. And team sports in America are even more bananas - you grow up supporting the Brooklyn Dodgers and a few years later they're 2.5k miles away with a new name. This seems a perfect example of what's happened / happening to the Republican Party - it's not the same party any more, but everyone who tied their entire personality to cheering for the red team is still cheering for it as it burns the country to the ground. I predict that inside ten years it will have also had the name change and probably be run out of Florida or somewhere.
https://www.cbsnews.com/news/kamala-harris-endorsement-bush-...
Trump caused a big political realignment actually.
I'm trans. this Administration does not like us. after Charlie Kirk's murder, things got legitimately scary. Musk was retweeting people who called us "deranged bioweapons" who needed to be "forcibly institutionalized." NSPM-7 is surveilling and infiltrating trans organizations. the Heritage Foundation proposed labeling us as "ideological extremists," in the same category as neo-Nazis. if I'm arrested, I'll go to a men's prison where I'll likely be given to a violent inmate as his cellmate to "pacify" him (V-coding.)
so yeah, I keep my head down. a lot of Jews kept their heads down in Germany in the '30s, you know? and just like then, it doesn't seem like other countries are too keen on taking us in as refugees. I hope that changes if things get bleak.
Wanna play at being a radical without popular support? Don't see the necessity to police the mentally ill and moderate the extremists in your group? Well, FAFO is all the rest of society can say.
Does being “extreme” justify extra-judicial violence?
This is not something to be proud of. You guys are giving yourself loaned freebies, retiring 5+ (!) years earlier than countries like BeNeLux and Germany, and are pretty much expecting the EU to eventually pick up the pieces which will drag us all down.
Edit: always lovely when HN downvotes truths :)
It just doesn't make sense to delay retirement while youth unemployment is such a big problem. We ALL should be fighting like France, in many aspects.
At some point France will be in too deep shit and will look to the EU to cover for them. We will all pay for that. And it is deeply unfair because other countries their citizens have accepted later retirement and more frugal benefits to keep their countries fiscally healthy.
France could cover the fiscal hole in other ways, but taxing corporations and wealth at a higher rate also consistently ends up being blocked. And each year the hole gets deeper.
Your theory doesn't actually match with reality, given that Macron's retirement reform was passed into law despite protests. As currently enacted, the age of retirement in France will progressively increase from 62 until reaching 64 in 2030.
Reform wasn't passed, it was forced via a technicality after riots made it politically unpalatable, and it has put France in a governing crisis ever since.
Also, retirement in North, West and Central EU is 67+, not 64. Greece is at 67 too, although begrudgingly.
Again, I'd be equally happy if France covers the fiscal hole some other way, but I am not going to cover for a country that is willingly becoming the sick man of Europe because they want to live comfortably on borrowed time. Which, by the way, is a literal repeat of Greece its crisis. Time is a flat circle indeed.
https://youtu.be/tMd7EfFsPIc (Video claims France is against them, but if they ever were they are not anymore)
- Control goes beyond politics
- A single, all-encompassing ideology
- No meaningful private sphere
- Mass mobilization and propaganda
- Extensive surveillance and repression
Seems like China is ticking all the boxes.
CIA/FBI have their own massive data centers (see snowden) inkl. their own older bigger palantr style software.
Elon Musk was able to connect a Starlink server to your data and no one cared. He and his Duche aeh sry doge baby boys were able to access and download all Social Security Numbers.
If someone knows were Putin and all the other world leaders are at any given moment, I would bet its USA first than China if even because i don't think China cares that much about it than USA does.
And everyone out of scope of this, lives probably in some rural USA town were no one cares for you at all anyway, but thats the same thing as in China.
- Control goes beyond politics
state corporation monopoly, 党支部 in private sector, crackdowns on NGOs and charities.
- A single, all-encompassing ideology
Party led, mandarin speaking Han Chinese nationalism, blended with Little Pink's unquestionable support for Xi and the party.
- No meaningful private sphere
社区网格员
- Mass mobilization and propaganda
We saw mobilizations on Chinese social media, attacking celebrities who don't openly say anything the party wants them to say. Mobilization in real life is rare though, cos it had shown it can backfire.
- Extensive surveillance and repression
Do I really need to explain this?
I'm in Hong Kong right now. Seems like it is still here to me.
Being self-righteous and a yank doesn't make sense, country of war mongers, something that cant be said of China.
> Covid saw people caged and sealed in their houses.
No. There were a few incidents very early on, when everyone was (quite understandably) panicking about a new, deadly virus that nobody had ever seen before, when some local city officials barred the doors of people who had just come from Wuhan. That was a scandal inside China, and it was immediately reversed.
What China did do quite extensively was border quarantine, and during localized outbreaks (caused by cases that slipped through quarantine at the border), mass testing and quarantine measures. This was during a once-in-a-generation pandemic that killed millions of people. In China, these measures saved several million lives. The estimates are that China's overall death rate was about 25% that of the US, and these measures are the reason. By the way, Taiwan and Australia took nearly identical measures, and I very much doubt that you would call them totalitarian societies.
Tell it to the people in Wuhan, and Shanghai, Urumqi, and other cities that had lockdowns. I was in Shanghai in 2022, I was confined to my apartment for nearly 3 months, you couldn't be more wrong.
Luckily laws still stand somewhat.
( And Trump ain't smart enough)
Have a peek at the fredom indx and the press freedom index for China. Guess where they stand?
You know about the chinese internet firewall.
You can't trust any data from the CCP.
And please don't equate the aberration that is the Trump administration with "regular" US administrations (and this is coming from a non US person).
But how free is the average North American, where getting sick can bring you and your family financial ruin? Where the "free press" is controlled by corporations who are also the main source of campaign funding for politicians? Where their urban spaces are designed to require you to have a car and promote complete atomized individuals?
Check out the Sean Ryan Show with Palmer Luckey on China and military tech.
Ridiculous take.
You do know that Chinese people do go to other countries and that we all can see how insanely racist they can be right?
No, China is not homogenous.
> racial problems are nonexistent
Ask a non-Han about how they feel about that statement.
(China wiped out the entire EU industry through a "quiet" trade war since like the last 15 years, and we're not really talking about that aren't we...)
The powers that be try to slow this down by banning imports outright (you can't for example import American chicken into Europe because of food safety laws), or high import taxes (Chinese EVs have a 50% import tax in Europe and the US to protect the local car manufacturers. Which is fair because the Chinese EV manufacturers are state-sponsored so their prices are unfair. Then again, western companies get billions in investor money to push the prices down).
EU/France has Mistral.
So again, stop referring to EU as a country, we are not, and it just annoys any Europeans as it comes of as "Americans who don't understand the world outside of the USA".
https://scrupulouspessimism.substack.com/p/america-means-the...
China’s governments actions are on a completely different level - for example:
“””
Since 2014, the government of the People's Republic of China has committed a series of ongoing human rights abuses against Uyghurs and other Turkic Muslim minorities in Xinjiang which has often been characterized as persecution or as genocide.
“”” https://en.wikipedia.org/wiki/Persecution_of_Uyghurs_in_Chin...
https://www.amnesty.org/en/location/asia-and-the-pacific/eas...
Yes Trump is clearly trying Totalitarianism in America, but it is orders of magnitude different from what is happening in China.
That should be at least comparable (if not worse) than what China is doing.
It's a little insane to me people comparing negatives of US and China. I mean, the simple fact we're allowed to say just about anything we want that is critical of the administration on this forum, in English and nothing happens is clear there is no comparison.
You have no idea the full breadth of the Chinese government because information is closed so quickly, in America it's all on display right in front.
China is repressing the Uyghur and threatening Taiwan. I don't agree with these actions but is really "orders of magnitude" worse than the destruction the US facilitates in the Middle East?
With Trump they are now openly hostile to European democracies, and ICE and doing their best at repression within the US.
> With Trump they are now openly hostile to European democracies, and ICE and doing their best at repression within the US.
And what is Europe going to do about it?
Boycott ChatGPT and Claude? Ha.
The point is US "soft power" is eroding incredibly rapidly and this will have consequences
by your logic gentrification of neighborhoods with different people moving in is genocide as well
Btw. remind me when last tiem China bombed school and killed 150+ school girls as your friend US?
Or as Brit I hope you are proud about all the killing your country participated in in illegal invasion to Iraq based on fake news about WMD.
Already do on EVs.
Really nice to see the Chinese are competing this strongly with the rest of the world. Competition is always nice for the end-consumer.
So does this mean I can run this on AMD? And on a consumer 9000 series card?
And in any case what does open source actually mean for an llm? It's not like you can look inside it to see what it's doing.
However there is so many factors involved beyond your control that it would not be a viable option compared to other possible security attacks.
It's like suggesting BYD has a high likelihood of making their cars into weapons or something. It's not in the company or their countries interest to do that.
Sure it could happen but I bet it would only happen in a targeted way. Why risk all credibility right now and engage in cyber warfare?
BYD and Tesla have the same ability to brick their cars anywhere. It's less a "weapon" and more a way to cripple a subset of people overnight if they so choose. A general major downside of "connected" products.
Meaning Tiktok in the us is complete garbage for kids, almost like a virus. Whereas in China it's more educational.
If I had to place a hidden target it'd probably be around RNGs or publicly exposed services..
I don't mean that flippantly. These things are dumped in the wild, used on common (largely) open source execution chains. If you find a software exploit, it's going to affect your population too.
Wet exploits are a bit harder to track. I'd assume there are plenty of biases based on training material but who knows if these models have a MKUltra training programme integrated into them?
Spearphishing.
Building reliance and exploiting it, through state subsidies, dumping, and market manipulation.
Handicapping provision to the west for competitive advantage.
Tech ceos are going around talking about how they will rule over employees and they will be unable to work in the future except for intelligence tokens. What if China commoditizes that without spending nearly as much resources? Kind of makes the trillions of dollars invested in the US a literal joke.
Of course there are risks.
I asked DS itself and it denied this. It says: 'Nvidia chips are absolutely used for DeepSeek V4. The reality is a pragmatic "both-and" strategy, not an "either-or."'
And based on the DS V4 technical report (https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...), it is mentioned that:
We validated the fine-grained EP scheme on both NVIDIA GPUs and HUAWEI Ascend NPUs platforms. Compared against strong non-fused baselines, it achieves 1.50 ~ 1.73× speedup for general inference workloads, and up to 1.96× for latency-sensitive scenarios such as RL rollouts and high-speed agent serving.
(In all honesty I relied on DS to give me the above, so I haven't vetted the information in full.)It mentions that Nvidia is still used. It doesn't even mention that Huawei chips are used in production — only in testing and validation, yes.
Bro, seriously?
Understandable.
In the west, especially in the USA, rich capitalists and warmongers control the narrative put forth in the news, which gets fed to the LLMs, which results in what you could call auto-censorship.
They manipulate the training data instead of censoring the model, but the result is the same.
> Summary: The U.S. is currently engaged in an active war aimed at dismantling the Iranian government and its military capabilities, but it distinguishes this from destroying the country or its people. However, the humanitarian impact—including civilian casualties from airstrikes and the domestic crackdown by Iranian security forces—has led many international observers to warn that the campaign risks long-term instability and "state collapse" rather than a simple transition of power.
It does do quite a bit better if you ask it about the genocide in Gaza, summarizing the case for it, and citing only token justifications from the guilty party.
As of April 2026, Gemini is... For very obvious reasons, highly biased towards cultural consensus. If your cultural consensus is strong on some really messed up things, that's the outcome that it's going to give you.
Irrespective of how close the outcomes are to the actual facts, those two things have a different quality, don't they?
Not as much a difference as you would wish, as mean of public discourse is very actively managed, to our collective detriment, by a very small group of powerful people, which often includes the government. It's the nature of mass media, and the incestuous relationship between power and reach.
They Thought They Were Free, and all that. By the time the 'mean of public discourse' centers on something incredibly stupid or awful, nobody can be arsed to figure out who planted that idea in our heads.
In reality it's only the terminally online that seem to create these narratives.
My point isn't to pick one side or the other, but agreeing with the other poster that the LLMs are not trained specifically to parrot administration propaganda.
I have a collection of novel probability and statistics problems at the masters and PhD level with varying degrees of feasibility. My test suite involves running these problems through first (often with about 2-6 papers for context) and then requesting a rigorous proof as followup. Since the problems are pretty tough, there is no quantitative measure of performance here, I'm just judging based on how useful the output is toward outlining a solution that would hopefully become publishable.
Just prior to this model, Gemini led the pack, with GPT-5 as a close second. No other model came anywhere near these two (no, not even Claude). Gemini would sometimes have incredible insight for some of the harder problems (insightful guesses on relevant procedures are often most useful in research), but both of them tend to struggle with outlining a concrete proof in a single followup prompt. This DeepSeek V4 Pro with max thinking does remarkably well here. I'm not seeing the same level of insights in the first response as Gemini (closer to GPT-5), but it often gets much better in the followup, and the proofs can be _very_ impressive; nearly complete in several cases.
Given that both Gemini and DeepSeek also seem to lead on token performance, I'm guessing that might play a role in their capacity for these types of problems. It's probably more a matter of just how far they can get in a sensible computational budget.
Despite what the benchmarks seem to show, this feels like a huge step up for open-weight models. Bravo to the DeepSeek team!
Summary: Opus 4.6 forms the baseline all three are trying to beat. DeepSeek V4-Pro roughly matches it across the board, Kimi K2.6 edges it on agentic/coding benchmarks, and Opus 4.7 surpasses it on nearly everything except web search.
DeepSeek V4-Pro Max shines in competitive coding benchmarks. However, it trails both Opus models on software engineering. Kimi K2.6 is remarkably competitive as an open-weight model. Its main weakness is in pure reasoning (GPQA, HMMT) where it trails Opus.
Speculation: The DeepSeek team wanted to come out with a model that surpassed proprietary ones. However, OpenAI dropped 5.4 and 5.5 and Anthropic released Opus 4.6 and 4.7. So they chose to just release V4 and iterate on it.
Basis for speculation? (i) The original reported timeline for the model was February. (ii) Their Hugging Face model card starts with "We present a preview version of DeepSeek-V4 series". (iii) V4 isn't multimodal yet (unlike the others) and their technical report states "We are also working on incorporating multimodal capabilities to our models."
Just ran a couple of them through GPT 5.5, but this is a single attempt, so take any of this with a grain of salt. I'm on the Plus tier with memory off so each chat should have no memory of any other attempt (same goes for other models too).
It seems to be getting more of the impressive insights that Gemini got and doing so much faster, but I'm having a really hard time getting it to spit out a proper lengthy proof in a single prompt, as it loves its "summaries". For the random matrix theory problems, it also doesn't seem to adhere to the notation used in the documents I give it, which is a bit weird. My general impression at the moment is that it is probably on par with Gemini for the important stuff, and both are a bit better than DeepSeek.
I can't stress how much better these three models are than everything else though (at least in my type of math problems). Claude can't get anything nontrivial on any of the problems within ten (!!) minutes of thinking, so I have to shut it off before I run into usage limits. I have colleagues who love using Claude for tiny lemmas and things, so your mileage may vary, but it seems pretty bad at the hard stuff. Kimi and GLM are so vague as to be useless.
- One problem on using quantum mechanics and C*-algebra techniques for non-Markovian stochastic processes. The interchange between the physics and probability languages often trips the models up, so pretty much everything tends to fail here.
- Three problems in random matrix theory and free probability; these require strong combinatorial skills and a good understanding of novel definitions, requiring multiple papers for context.
- One problem in saddle-point approximation; I've just recently put together a manuscript for this one with a masters student, so it isn't trivial either, but does not require as much insight.
- One problem pertaining to bounds on integral probability metrics for time-series modelling.
Happy to try to answer more specific questions if anyone has any, but yes, these are among my active research projects so there's only so much I can say.
https://api-docs.deepseek.com/guides/thinking_mode
No BS, just a concise description of exactly what I need to write my own agent.
Western Models are optimizing to be used as an interchangeable product. Chinese models are being optimizing to be built upon.
But so much investment in their platforms, not just their APIs?
Why? It sounds like the stupidest idea ever. Interchangeability = no lock-in = no moot.
Now that you’re winning, others start cloning your API to siphon your users.
Now that you’re losing, you start cloning the current winner, who is probably a clone of your clone.
Highly competitive markets tend to normalize, because lock-in is a cost you can’t charge and remain competitive. The customer holds power here, not the supplier.
Thats also why everyone is trying to build into the less competitive spaces, where they could potentially moat. Tooling, certs, specialized training data, etc
They are developing their moats with the platform tooling around it right now though. Look at Anthropic with Routines and OpenAI with Agents. Drop that capability in to a business with loose controls and suddenly you have a very sticky product with high switching costs. Meanwhile if you stick with purely the ‘chat’ use cases, even Cowork and scheduled tasks, you maintain portability.
In my tests too[0], it doesn't reach top 10. One issue, which they also mentioned in their post, is that they can't really serve well the model at the moment, so V4-Pro is heavily rate-limited and gives a lot of timeout errors when I try to test it. This shouldn't be an issue though, considering the model is open-source, but it makes it hard to accurately test at the moment.
[0]: https://aibenchy.com/compare/deepseek-deepseek-v4-flash-high...
Pretty cool, I think they're the first to guarantee determinism with the fixed seed or at the temperature 0. Google came close but never guaranteed it AFAIK. DeepSeek show their roots - it may not strictly be a SotA model, but there's a ton of low-level optimizations nobody else pays attention to.
"Limited by the capacity of high-end computational resources, the current throughput of the Pro model remains constrained. We expect its pricing to decrease significantly once the Ascend 950 has been deployed into production."
https://api-docs.deepseek.com/zh-cn/news/news260424#api-%E8%...
https://api-docs.deepseek.com/zh-cn/news/news260424#api-%E8%...
This is the first figure of the section that the above links point to (https://api-docs.deepseek.com/zh-cn/img/v4-spec.png).
And I can read Chinese.
I’d like somebody to explain to me how the endless comments of "bleeding edge labs are subsidizing the inference at an insane rate" make sense in light of a humongous model like v4 pro being $4 per 1M. I’d bet even the subscriptions are profitable, much less the API prices.
edit: $1.74/M input $3.48/M output on OpenRouter
It is like car vs. kick scooter.
I think you need to define "can get coding work done" for this to make sense. Ive been using GPT-3 back-then for basic scripts, does that count ? Or only Claude-Code ?
I also think this is a false dichotomy, if you look at the Project Vend project or Vending-Bench, customer support etc. is at no means trivial. (Old but great story https://www.businessinsider.com/car-dealership-chevrolet-cha...)
What im really hoping is for a double-punch like with V3 -> R1
In 2023, the depreciation schedule for H100s was 2 years, but they are still oversubscribed and generating signficant income.
Coreweve has upped their depreciation for GPUs to 6 years(!) now, which seems more realistic.
https://www.silicondata.com/blog/h100-rental-price-over-time
We therefore cannot just look at inference costs directly, training is part of the pitch. Without the promises of continuous improvement and chasing the elusive AGI, money for investments for inference evaporates.
And Microsoft are going the same route to moving Copilot Cowork over to a utilisation based billing model which is very unusual for their per seat products (I’m actually not sure I can ever remember that happening).
But seriously, it just stems from the fact some people want AI to go away. If you set your conclusion first, you can very easily derive any premise. AI must go away -> AI must be a bad business -> AI must be losing money.
There are still major unanswered questions here. For instance, all of the incremental data capacity build out is going to businesses that have totally unknown LT unit economics and that today are burning obscene amounts of cash.
Aka: everyone who uses Nvidia isn't selling at cost, because Nvidia is so expensive.
At some point (from the very beginning till ~2025Q4) Claude Code's usage limit was so generous that you can get roughly $10~20 (API-price-equivalent) worth of usage out of a $20/mo Pro plan each day (2 * 5h window) - and for good reason, because LLM agentic coding is extremely token-heavy, people simply wouldn't return to Claude Code for the second time if provided usage wasn't generous or every prompt costs you $1. And then Codex started trying to poach Claude Code users by offering even greater limits and constantly resetting everyone's limit in recent months. The API price would have to be 30x operating cost to make this not a subsidy. That would be an extraordinary claim.
eg:
Token prices are significantly subsidized and anyone that does any serious work with AI can tell you this.
https://news.ycombinator.com/item?id=47684887
(the claims don't make any sense, but they are widely held)
I think I understand the major reasons for this meme, but I find it really worrying; there were lots of incorrect ‘it’s a bubble’ conversations here in 2012-2015, but I don’t think they had the pervasive nature and “obvious” conclusion that a whole generation of engineering talent should just, you know, leave.
Meanwhile I am hearing rational economic modeling from the companies selling inference; Jensen, (a polished promoter, I grant you) says it really well — token value is increasing radically, in that new models -> better quality, and therefore revenues and utilization are increasing, and therefore contrary to the popular financial and techbro modeling of 2023, things like A100s still cost quite a lot whether hourly or to purchase. (!) Basically the economic value is so strong that it has actually radically extended the life of hardware.
I just hate to imagine like half of the world’s (or US’s) engineering talent quitting, spending ten years afraid, or wrongly convinced of some ‘inevitable’ market outcome. Feels like it will be bad for people’s personal lives, and bad for progress simultaneously.
I'm still playing with the new Qwen3.6 35B and impressed, now DeepSeek v4 drops; with both base and instruction-tuned weights? There goes my weekend :P
One answer - Chinese Communist Party. They are being subsidized by the state.
Edit: it seems "open source" was edited out of the parent comment.
no one is ever going to release their training data because it contains every copyrighted work in existence. everyone, even the hecking-wholesome safety-first Anthropic, is using copyrighted data without permission to train their models. there you go.
It is very much a valuable thing already, no need to taint it with wrong promise.
Though I disagree about being used if it was indeed open source: I might not do it inside my home lab today, but at least Qwen and DeepSeek would use and build on what eg. Facebook was doing with Llama, and they might be pushing the open weights model frontier forward faster.
They're both correct given how the terms are actually used. We just have to deduce what's meant from context.
There was a moment, around when Llama was first being released, when the semantics hadn't yet set. The nutter wing of the FOSS community, to my memory, put forward a hard-line and unworkable definition of open source and seemed to reject open weights, too. So the definition got punted to the closest thing at hand, which was open weights with limited (unfortunately, not no) use restrictions. At this point, it's a personal preference that's at most polite to respect if you know your audience has one.
https://www.reuters.com/technology/nvidia-is-sued-by-authors...
The training scripts are in Megatron and vLLM.
1. Training data is the source. 2. Training is compilation/compression. 3. Weights are the compiled source akin to optimized assembly.
However it's an imperfect analogy on so many levels. Nitpick away.
For reference, the huawei Ascend 950 that this thing runs on is supposed to be roughly comparable to nVidia's H100 from 2022. In other words, things are hotting up in the GPU war!
Nvidia's forward PE ratio is only 20 for 2026. That's much lower than companies like Walmart and Costco. It's also growing nearly 100% YoY and has a $1 trillion backlog.
I think Nvidia is cheap.
That's a very strange comment. Why would anyone run a dense model on a low-end computer? A 8B model is only going to make sense if you have a dGPU. And a Qwen3.6 or Gemma4 MoE aren't going to be “beaten the hell out” for most tasks especially if you use tools.
Finally, over the lifetime of your computer, your ChatGPT subscription is going to cost more than the cost of your reference computer! So the real question should be whether you're better off with a $1000 computer and a ChatGPT subscription or with a $2000 computer (assuming a conservative lifetime of 4 years for the computer).
My Strix Halo desktop (which I paid ~1700€ before OpenAI derailed the RAM market) paired with Qwen3.5 is a close replacement for a $200/month subscription, so the cost/benefit ratio is strongly in favor of the local model in my use case.
The complexity of following model releases and installing things needed for self-hosting is a valid argument against local models, but it's absolutely not the same thing as saying that local models are too bad to use (which is complete BS).
Biggest risk I see is Nvidia having delays / bad luck with R&D / meh generations for long enough to depress their growth projections; and then everything gets revalued.
At this point I would just pick the one who's "ethics" and user experience you prefer. The difference in performance between these releases has had no impact on the meaningful work one can do with them, unless perhaps they are on the fringes in some domain.
Personally I am trying out the open models cloud hosted, since I am not interested in being rug pulled by the big two providers. They have come a long way, and for all the work I actually trust to an LLM they seem to be sufficient.
New model comes out, has some nice benchmarks, but the subjective experience of actually using it stays the same. Nothing's really blown my mind since.
Feels like the field has stagnated to a point where only the enthusiasts care.
Since then it's just been a cycle of the old model being progressively lobotomised and a "new" one coming out that if you're lucky might be as good as the OG Opus 4.5 for a couple of weeks.
Subjective but as far as I can tell no progress in almost a year, which is a lifetime in 2022-25 LLM timelines
Back in Nov 2025, Opus 4.5 (80.9%) was the first proprietary model to do so.
So it os hard to tell how much of a model gain is due to skill, and how much - overfitting.
I just want to remind you that this is happening at the same time as Anthropic A/B tests removal of Code from Pro Plan, and as OpenAI releases gpt-5.5 2x more expensive than gpt-5.4...
That’s a big if. It’s my experience that models that perform very well on benchmarks do not necessarily perform well in real life.
I’ve mostly started ignoring the benchmarks and run my own evals.
`https://openrouter.ai/api/messages with model=deepseek/deepseek-v4-pro, OR returns an error because their Anthropic-compat translator doesn't cover V4 yet. The Claude CLI dutifully surfaces that error as "model...does not exist"
Stuff that was prohibitive six months ago is now up for grabs. We keep on working on the infra level now, swithcing models whenever we run out of credits, or want a different result. The question is how do we build context, architecture and ensure the agent is effective and efficient..... wouldn't it be good if we simply used less energy to make these AI calls?
And we got new base models, wonderful, truly wonderful
But more broadly: openrouter solves the problem of making a broad range of models available with a single payment endpoint, so you can just switch around as much as you like.
Another way to keep the ability to try out new models is to buy a reseller subscription like Cursor’s.
For OSS model, I have z.ai yearly subscription during the promo. But it's a lot more expensive now. The model is good imo, and just need to find the right providers. There are a lot of alternatives now. Like I saw some good reviews regarding ollama cloud.
If you're trying to make a buck while unemployed, sure get a subscription. Otherwise learn how to work again without AI, just focus on the interesting stuff.
I'm on Max x5 plan and any of the 'good' models like Kimi 2.6, GLM, DeepSeek would have cost 3-5x in per-token billing for what I used on my Claude plan the last three months
So unless my Claude fudged the maths to make itself look better, seems like I'm getting a good deal
This “no harm to me” meme about a foreign totalitarian government (with plenty of incentive to run influence ops on foreigners) hoovering your data is just so mind-bogglingly naive.
Relatively speaking, DeepSeek is less untrustworthy than Grok.
When I try ChatGPT on current events from the White House it interprets them as strange hypotheticals rather than news, which is probably more a problem with DC than with GPT, but whatever.
That would be a great argument if the American models weren’t so heavily censored.
The Chinese model might dodge a question if I ask it about 1-2 specific Chinese cultural issues but then it also doesn ’t moralize me at every turn because I asked it to use a piece of security software.
Even for minor stuff like beeing addicted to drugs.
Looks pretty totalitarian to me.
Note: you can have this conversation criticizing the US on a US website. Try criticizing Xi or the CCP or calling him Pooh on a Chinese website.
You think China doesn’t imprison drug users?
China recently executed a low level drug trafficker
https://www.lemonde.fr/en/international/article/2026/04/05/c...
China is one of the top executioners. China executes more than rest of the world combined
https://www.amnesty.org/en/latest/news/2017/04/china-must-co...
You think China is honest about political prisoners in Tibet and Xinjiang?
Criticize the US all you want but I can’t understand the whitewashing of a real totalitarian and genocidal state like mainland China.
But if we start nitpicking the US also executes people all over the world without trial and has secret prisons worldwide where they put people (guess what) without trial.
Not quite the same.
Quick google top link
https://en.wikipedia.org/wiki/Forced_organ_harvesting_from_F...
yes, this is exactly what I'm saying.
This is why I’ve been urging everyone I know to move away from American based services and providers. It’s slow but honest work.
But for folks on the opposite side of the world, the threats are more like "they're selling us electric cars and solar panels too cheaply" and the hypothetical "these super cheap CCTV cameras could be used for remote spying"
China is a nation built for peace, while western nations are built for war.
The US is (mostly) protective of its citizens but (depending on administration) varyingly hostile to outsiders (immigrants, starting wars, etc.).
China is suppressive towards its own citizens, but has been largely peaceful with other countries and immigrants/visitors. (Granted, China has way fewer immigrants than the US, so this is not comparable).
Feel free to go post similar on Chinese social media about their leaders.
The executive branch?
Half the country would be locked up right now if they weren’t allowed to criticize Trump. Have you even paid attention to how much he’s shitted on, on a daily basis?
My country’s per capita income is $2500 a year. We can’t pay perpetual rent to OAI/Anthropic
This sounds whole lot like potatoh potahto. I think the former argument is very much the correct one: China can undercut everyone and win, even at a loss. Happened with solar panels, steel, evs, sea food - it's a well tested strategy and it works really well despite the many flavors it comes in.
That being said a job well done for the wrong reasons is still a job well done so we should very much welcome these contributions, and maybe it's good to upset western big tech a bit so it's remains competitive.
The decisions to mobilize a large rural base toward manufacturing and the central bank goals to keep the yuan cheap as a critical support of this project were absolutely national.
They were ultimately about bringing (or trying to bring) one of the most populous nations in the world out of extreme poverty; in particular the people of the country out of extreme poverty.
There are different policies in place today, and, crucially, bleeding edge tech is not gainful labor employment —- BYD has some factories with roughly 2 employees per acre of robotic production, for instance. Or datacenters where the revenue could scale but the labor will not.
So, these are different times, different goals, different political and labor outcomes. Reasoning about what China “must do”, or has as a matter of “national policy” should start with a clear look at history and circumstance, or you’re likely to read things incorrectly.
Just this week they published a serious foundational library for LLMs https://github.com/deepseek-ai/TileKernels
Others worth mentioning:
https://github.com/deepseek-ai/DeepGEMM a competitive foundational library
https://github.com/deepseek-ai/Engram
https://github.com/deepseek-ai/DeepSeek-V3
https://github.com/deepseek-ai/DeepSeek-R1
https://github.com/deepseek-ai/DeepSeek-OCR-2
They have 33 repos and counting: https://github.com/orgs/deepseek-ai/repositories?type=all
And DeepSeek often has very cool new approaches to AI copied by the rest. Many others copied their tech. And some of those have 10x or 100x the GPU training budget and that's their moat to stay competitive.
The models from Chinese Big Tech and some of the small ones are open weights only. (and allegedly benchmaxxed) (see https://xcancel.com/N8Programs/status/2044408755790508113). Not the same.
So you can’t see what facts are pruned out, what biases were applied, etc. Even more importantly, you can’t make a slightly improved version.
This model is as open source as a windows XP installation ISO.
Did you even read my comment?
And you think the US tech giants don't have any ulterior motives?!
Model was released and it's amazing. Frontier level (better than Opus 4.6) at a fraction of the cost.
As a non-Opus user, I'll continue to use the cheapest fastest models that get my job done, which (for me anyway) is still MiniMax M2.5. I occasionally try a newer, more expensive model, and I get the same results. I have a feeling we might all be getting swindled by the whole AI industry with benchmarks that just make it look like everything's improving.
The tricky part is that the "number of tokens to good result" does absolutely vary, and you need a decent harness to make it work without too much manual intervention, so figuring out which model is most cost-effective for which tasks is becoming increasingly hard, but several are cost-effective enough.
Codex is just so much better, or the genera GPT models.
https://github.blog/news-insights/company-news/changes-to-gi...
Substantially worse at following instructions and overoptimized for maximizing token usage
I do some stuff with gemini flash and Aider, but mostly because I want to avoid locking myself into a walled garden of models, UIs and company
If you're feeling frisky, Zed has a decent agent harness and a very good editor.
Opencode was getting there, but it seems the founders lost interest. Pi could be it, but its very focused on OpenClaw. Even Codex cli doesnt have all of it.
which harness works well with Deepseek v4 ?
So while I agree mixed model is the way to go, opus is still my workhorse.
Not saying it is better or worse, but the way I perpersonally prefer is to design in chat, to make sure all unknown unknown are addressed
In contrast ChatGPT 5.3 and also Opus has a 90% rate at least on this same project. (Embedded)
All other tests were the same. What are you doing with these models?
This is free... as in you can download it, run it on your systems and finetune it to be the way you want it to be.
In theory, sure, but as other have pointed out you need to spend half a million on GPUs just to get enough VRAM to fit a single instance of the model. And you’d better make sure your use case makes full 24/7 use of all that rapidly-depreciating hardware you just spent all your money on, otherwise your actual cost per token will be much higher than you think.
In practice you will get better value from just buying tokens from a third party whose business is hosting open weight models as efficiently as possible and who make full use of their hardware. Even with the small margin they charge on top you will still come out ahead.
And that GPU wouldn’t run one instance, the models are highly parallelizable. It would likely support 10-15 users at once, if a company oversubscribed 10:1 that GPU supports ~100 seats. Amortized over a couple years the costs are competitive.
Obviously, and certainly companies do run their own models because they place some value on data sovereignty for regulatory or compliance or other reasons. (Although the framing that Anthropic or OpenAI might "steal their data" is a bit alarmist - plenty of companies, including some with _highly_ sensitive data, have contracts with Anthropic or OpenAI that say they can't train future models on the data they send them and are perfectly happy to send data to Claude. You may think they're stupid to do that, but that's just your opinion.)
> the models are highly parallelizable. It would likely support 10-15 users at once.
Yes, I know that; I understand LLM internals pretty well. One instance of the model in the sense of one set of weights loaded across X number of GPUs; of course you can then run batch inference on those weights, up to the limits of GPU bandwidth and compute.
But are those 100 users you have on your own GPUs usings the GPUs evenly across the 24 hours of the day, or are they only using them during 9-5 in some timezone? If so, you're leaving your expensive hardware idle for 2/3 of the day and the third party providers hosting open weight models will still beat you on costs, even without getting into other factors like they bought their GPUs cheaper than you did. Do the math if you don't believe me.
Now, at the moment, i can still use 4.6 but eventually Anthropic are going to remove it, and when it's gone it will be gone forever. I'm planning on trying Deepseek v4, because even if it's not quite as good, I know that it will be available forever, I'll always be able to find someone to run it.
Already these models are useful for a myriad of use cases. It's really not that important if a model can 1-shot a particular problem or draw a cuter pelican on a bike. Past a degree of quality, process and reliability are so much more important for anything other than complete hands-off usage, which in business it's not something you're really going to do.
The fact that my tool may be gone tomorrow, and this actually has happened before, with no guarantees of a proper substitute... that's a lot more of a concern than a point extra in some benchmark.
If you want to go budget corporate, 7 x H200 is just barely going to run it, but all in, $300k ought to do it.
- To run at full precision: "16–24 H100s", giving us ~$400-600k upfront, or $8-12/h from [us-east-1](https://intuitionlabs.ai/articles/h100-rental-prices-cloud-c...).
- To run with "heavy quantization" (16 bits -> 8): "8xH100", giving us $200K upfront and $4/h.
- To run truly "locally"--i.e. in a house instead of a data center--you'd need four 4090s, one of the most powerful consumer GPUs available. Even that would clock in around $15k for the cards alone and ~$0.22/h for the electricity (in the US).
Truly an insane industry. This is a good reminder of why datacenter capex from since 2023 has eclipsed the Manhattan Project, the Apollo program, and the US interstate system combined...
10 years from now that hardware will be on eBay for any geek with a couple thousand dollars and enough power to run it.
"671B total / 37B active"
"Full precision (BF16)"
And they claim they ran this non-existent model on vLLM and SGLang over a month and a half ago.
It's clickbait keyword slop filled in with V3 specs. Most of the web is slop like this now. Sigh.
It's about 2 months behind GPT 5.5 and Opus 4.7.
As long as it is cheap to run for the hosting providers and it is frontier level, it is a very competitive model and impressive against the others. I give it 2 years maximum for consumer hardware to run models that are 500B - 800B quantized on their machines.
It should be obvious now why Anthropic really doesn't want you to run local models on your machine.
Doesn't mean Deepseek v4 isn't great, just benchmarks alone aren't enough to tell.
> In our internal evaluation, DeepSeek-V4-Pro-Max outperforms Claude Sonnet 4.5 and approaches the level of Opus 4.5.
If its coding abilities are better than Claude Code with Opus 4.6 then I will definitely be switching to this model.
It's still a "preview" version atm.
I don't see why Deepseek would care to respect Anthropic's ToS, even if just to pretend. It's not like Anthropic could file and win a lawsuit in China, nor would the US likely ban Deepseek. And even if the US gov would've considered it, Anthropic is on their shitlist.
There we go again :) It seems we have a release each day claiming that. What's weird is that even deepseek doesn't claim it's better than opus w/ thinking. No idea why you'd say that but anyway.
Dsv3 was a good model. Not benchmaxxed at all, it was pretty stable where it was. Did well on tasks that were ood for benchmarks, even if it was behind SotA.
This seems to be similar. Behind SotA, but not by much, and at a much lower price. The big one is being served (by ds themselves now, more providers will come and we'll see the median price) at 1.74$ in / 3.48$ out / 0.14$ cache. Really cheap for what it offers.
The small one is at 0.14$ in / 0.28$ out / 0.028$ cache, which is pretty much "too cheap to matter". This will be what people can run realistically "at home", and should be a contender for things like haiku/gemini-flash, if it can deliver at those levels.
LMAO
I have no idea why you'd think that, but this is straight from their announcement here (https://mp.weixin.qq.com/s/8bxXqS2R8Fx5-1TLDBiEDg):
> According to evaluation feedback, its user experience is better than Sonnet 4.5, and its delivery quality is close to Opus 4.6's non-thinking mode, but there is still a certain gap compared to Opus 4.6's thinking mode.
This is the model creators saying it, not me.
Claude4.6 was almost 10pp better at at answering questions from long contexts ("corpuses" in CorpusQA and "multiround conversations" in MRCR), while DSv4 was a staggering 14pp better at one math challenge (IMOAnswerBench) and 12pp better at basic Q&A (SimpleQA-Verified).
That's literally what the I Ching calls "good fortune."
Competition, when no single dragon monopolizes the sky, brings fortune for all.
input: $0.14/$0.28 (whereas gemini $0.5/$3)
Does anyone know why output prices have such a big gap?
This is a pretty interesting thing they've built in my opinion, and not something I'd expect to be buried in the model paper like this. Does anyone have any details about it? Google doesn't seem to find anything of note, and I'd love to dive a bit deeper into DSec.
The website now has a link to the announcement on Twitter here https://x.com/deepseek_ai/status/2047516922263285776
Copying text of that below
DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length.
DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models.
DeepSeek-V4-Flash: 284B total / 13B active params. Your fast, efficient, and economical choice.
Try it now at http://chat.deepseek.com via Expert Mode / Instant Mode. API is updated & available today!
Tech Report: https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...
Open Weights: https://huggingface.co/collections/deepseek-ai/deepseek-v4
https://api-docs.deepseek.com/guides/coding_agents#integrate...
But in this case, it's more likely just to be a tooling issue.
https://simonwillison.net/2026/Apr/24/deepseek-v4/
Both generated using OpenRouter.
For comparison, here's what I got from DeepSeek 3.2 back in December: https://simonwillison.net/2025/Dec/1/deepseek-v32/
And DeepSeek 3.1 in August: https://simonwillison.net/2025/Aug/22/deepseek-31/
And DeepSeek v3-0324 in March last year: https://simonwillison.net/2025/Mar/24/deepseek/
As in have the model consider its generated SVG, and gradually refine it, using its knowledge of the relative positions and proportions of the shapes generated, and have it spin for a while, and hopefully the end result will be better than just oneshotting it.
Or maybe going even one step further - most modern models have tool use and image recognition capabilities - what if you have it generate an SVG (or parts/layers of it, as per the model's discretion) and feed it back to itself via image recognition, and then improve on the result.
I think it'd be interesting to see, as for a lot of models, their oneshot capability in coding is not necessarily corellated with their in-harness ability, the latter which really matters.
I should try it again with the more recent models.
Could you please try with Opus 4.7? I think there's a chance of it doing well, considering the design/vision focus.
Let me tell you how much the Pro one sucks... It looks like failed Pedersen[1]. The rear wheel intersects with the bottom bracket, so it wouldn't even roll. Or rather, this bike couldn't exist.
The flash one looks surprisingly correct with some wild fork offset and the slackest of seat tubes. It's got some lowrider[2] aspirations with the small wheels, but with longer, Rivendellish[3], chainstays. The seat post has different angle than the seat tube, so good luck lowering that.
[1] https://en.wikipedia.org/wiki/Pedersen_bicycle
I wonder which model will try some more common spoke lacing patterns. Right now there seems to be a preference for radial lacing, which is not super common (but simple to draw). The Flash and Pro one uses 16 spoke rims, which actually exist[1] but are not super common.
The Pro model fails badly at the spokes. Heck, the spokes sit on the outside of the drive side of the rim and tire. Have a nice ride riding on the spokes (instead of the tire) welded to the side of your rim.
Both bikes have the drive side on the left, which is very very uncommon. That can't exist in the training data.
[1] https://cicli-berlinetta.com/product/campagnolo-shamal-16-sp...
at the top of the linked pages.
1) LLM is not AGI. Because surely if AGI it would imply that pro would do better than flash?
2) and because of the above, Pelican example is most likely already being benchmaxxed.
How much does the drawing change if you ask it again?
Gemini-3.1-Pro at 91.0
Opus-4.6 at 89.1
GPT-5.4, Kimi2.6, and DS-V4-Pro tied at 87.5
Pretty impressive
If AI was so good at coding, why can’t it actually make a usable Gemini/AI Studio app?
In my experience, Gemini is the most insightful model for hard problems (particularly math problems that I work on).
For context, for an agent we're working on, we're using 5-mini, which is $2/1m tokens. This is $0.30/1m tokens. And it's Opus 4.6 level - this can't be real.
I am uncomfortable about sending user data which may contain PII to their servers in China so I won't be using this as appealing as it sounds. I need this to come to a US-hosted environment at an equivalent price.
Hosting this on my own + renting GPUs is much more expensive than DeepSeek's quoted price, so not an option.
As a European I feel deeply uncomfortable about sending data to US companies where I know for sure that the government has access to it.
I also feel uncomfortable sending it to China.
If you'd asked me ten years ago which one made me more uncomfortable. China.
But now I'm not so sure, in fact I'm starting to lean towards the US as being the major risk.
It's doesn't seem all that out there compared to the other Chinese model price/performance? Kimi2.6 is cheaper even than this, and is pretty close in performance
Which strikes me as odd - Inwoukd have assumed someone had an edge in terms of at least 10% extra GPUs.
"Not seduced by praise, not terrified by slander; following the Way in one's conduct, and rectifying oneself with dignity." (不诱于誉,不恐于诽,率道而行,端然正己)
(It is mainly used to express the way a Confucian gentleman conducts himself in the world. It reminds me of an interview I once watched with an American politician, who said that, at its core, China is still governed through a Confucian meritocratic elite system. It seems some things have never really changed.
In some respects, Liang Wenfeng can be compared to Linux. The political parallel here is that the advantages of rational authoritarianism are often overlooked because of the constraints imposed by modern democratic systems. )
Not gonna happen
dang, probably the two should be merged and that be the link
Codex shows ~258k for me and Claude Code often shows ~200k, so I’m curious how DeepSeek is exposing such a large window.
The 1M window might be usable, but it will probably underperform against a smaller window of course.
A mac with 256 GB memory would run it but be very slow, and so would be a 256GB ram + cheapo GPU desktop, unless you leave it running overnight.
The big model? Forget it, not this decade. You can theoretically load from SSD but waiting for the reply will be a religious experience.
Realistically the biggest models you can run on local-as-in-worth-buying-as-a-person hardware are between 120B and 200B, depending on how far you’re willing to go on quantization. Even this is fairly expensive, and that’s before RAM went to the moon.
Strix halo has 256 GB/s bandwidth for $2500. The Flash model has 13 GB activations.
256 / 13 = 19.6 tokens per second
Except you cannot fit it into the maximum RAM of 128 GB Strix Halo supports. So move on.
Another option is Threadripper. That's 8 memory channels. Using older DDR4-3200 you get roughly 200 GB/s. For $2000.
200 / 13 = 15.4 tokens per second
But, a chunk of per-token weights is actually always the same and not MoE, so you would offload that to a GPU and get a decent speedup. Say 25 tokens per second total.
Then likely some expensive Mac. No idea.
Eventually you arrive at a mining rig chassis with a beefy board and multiple GPUs. That has the benefit of pipelining. You run part of the model on one GPU and move on, so another batch can start on the first one. Low (say 30-100) tps individually, but a lot more in parallel. Best get it with other people.
The flash version here is 284B A13B, so it might perform OK with a fairly small amount of VRAM for the active params and all regular ram for the other params, but I’d have to see benchmarks. If it turns out that works alright, an eBay server plus a 3090 might be the bang-for-buck champ for about $2.5K (assuming you’re starting from zero).
"Due to constraints in high-end compute capacity, the current service capacity for Pro is very limited. After the 950 supernodes are launched at scale in the second half of this year, the price of Pro is expected to be reduced significantly."
So it's going to be even cheaper
But if it does, then in the following week we'll see DeepSeek4 floods every AI-related online space. Thousands of posts swearing how it's better than the latest models OpenAI/Anthropic/Google have but only costs pennies.
Then a few weeks later it'll be forgotten by most.
If one finds it difficult to set up OpenCode to use whatever providers they want, I won't call them 'dev'.
The only real friction (if the model is actually as good as SOTA) is to convince your employer to pay for it. But again if it really provides the same value at a fraction of the cost, it'll eventually cease to be an issue.
"If one finds it difficult to set up OpenCode to use whatever providers they want, I won't call them 'dev'."
I feel the same way. But look at the ollama vs llama.cpp post from HN few days back and you will see most of the enthusiasts in this space are very non technical people.Was expecting that the release would be this month [1], since everyone forgot about it and not reading the papers they were releasing and 7 days later here we have it.
One of the key points of this model to look at is the optimization that DeepSeek made with the residual design of the neural network architecture of the LLM, which is manifold-constrained hyper-connections (mHC) which is from this paper [2], which makes this possible to efficiently train it, especially with its hybrid attention mechanism designed for this.
There was not that much discussion around it some months ago here [3] about it but again this is a recommended read of the paper.
I wouldn't trust the benchmarks directly, but would wait for others to try it for themselves to see if it matches the performance of frontier models.
Either way, this is why Anthropic wants to ban open weight models and I cannot wait for the quantized versions to release momentarily.
[0] https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...
[1] https://news.ycombinator.com/item?id=47793880
Do you have a source?
bash({"command":"gh pr create --title "Improve Calendar module docs and clean up idiomatic Elixir" --body "$(cat <<'EOF'
Problem
The Calendar modu...
like generating output, but not actually running the bash command so not creating the PR ultimately. I wonder if it's a model thing, or an opencode thing.> We present a preview version of DeepSeek-V4 series, including two strong Mixture-of-Experts (MoE) language models — DeepSeek-V4-Pro with 1.6T parameters (49B activated) and DeepSeek-V4-Flash with 284B parameters (13B activated) — both supporting a context length of one million tokens. DeepSeek-V4 series incorporate several key upgrades in architecture and optimization: (1) a hybrid attention architecture that combines Compressed Sparse Attention (CSA) and Heavily Compressed Attention (HCA) to improve long-context efficiency; (2) Manifold-Constrained Hyper-Connections (mHC) that enhance conventional residual connections; (3) and the Muon optimizer for faster convergence and greater training stability. We pre-train both models on more than 32T diverse and high-quality tokens, followed by a comprehensive post-training pipeline that unlocks and further enhances their capabilities. DeepSeek-V4-Pro-Max, the maximum reasoning effort mode of DeepSeek-V4-Pro, redefines the state-of-the-art for open models, outperforming its predecessors in core tasks. Meanwhile, DeepSeek-V4 series are highly efficient in long-context scenarios. In the one-million-token context setting, DeepSeek-V4-Pro requires only 27% of single-token inference FLOPs and 10% of KV cache compared with DeepSeek-V3.2. This enables us to routinely support one-million-token contexts, thereby making long-horizon tasks and further test-time scaling more feasible. The model checkpoints are available at https://huggingface.co/collections/deepseek-ai/deepseek-v4.
1: https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...
Keep an eye on https://huggingface.co/unsloth/models
Update ten minutes later: https://huggingface.co/unsloth/DeepSeek-V4-Pro just appeared but doesn't have files in yet, so they are clearly awake and pushing updates.
I have never tried one yet but I am considering trying that for a medium sized model.
As I understand it if DeepSeek v4 Pro is a 1.6T, 49B active that means you'd need just 49B in memory, so ~100GB at 16 bit or ~50GB at 8bit quantized.
v4 Flash is 284B, 13B active so might even fit in <32GB.
V4 is natively mixed FP4 and FP8, so significantly less than that. 50 GB max unquantized.
My Mac can fit almost 70B (Q3_K_M) in memory at once, so I really need to try this out soon at maybe Q5-ish.
Streaming weights from RAM to GPU for decode makes no sense at all because batching requires multiple parallel streams.
Streaming weights from SSD _never_ makes sense because the delta between SSD and RAM is too large. There is no situation where you would not be able to fit a model in RAM and also have useful speeds from SSD.
Note: these were just two that I starred when I saw them posted here. I have not looked seriously at it at the moment,