upvote
This is sort of what their first sentence states? Except your line implies that they are fast in training and inference, they imply they are focusing on inference and are dropping training speed for it.

It's a nice opening as it is imo

reply
They don't say anything about dropping training speed.
reply
> a departure from Mamba-2, which optimized for training speed.

?

reply
The first sentence basically does though, no?
reply
Of course my only objection was the language. LLMs are now old enough to leave the jargon behind and talk in simple easy to understand terms.
reply
I’d argue the opposite, the terminology is fairly mainstream by now and “inference” has a much more specific sense than “making predictions”.
reply
The blog is technical, technical terms in the TL;DR seems relevant to me.
reply
I don't get the downvotes, as I had trouble understanding the intro as well. It seems it was written for a very specific audience.
reply
Yes, it is written for a specific audience.

That is not a reason for snark.

As other commenters have noted, it’s well written.

reply
> I don't get the downvotes

Because the blog post is a technical one and the intro contains very common jargon, and the proposed alternative was wrong.

reply
I don’t know why you’re being downvoted. As a longtime editor your version is immensely better. Looks like the original was probably not human-written.
reply