upvote
> ended up enabling groq

For those reading fast, this isn't a reference to SpaceX's Grok, this is Groq.com - with its custom inference chip, and offerings like https://groq.com/blog/introducing-llama-3-groq-tool-use-mode... and https://console.groq.com/landing/llama-api

reply
Really liked Groq due to its speed but it seems like after Nvidia bought it it has been discontinued...
reply
> He then open sourced llama and wanted to be the android of llms.

Well the original llama did kick off the era of open source LLMs. Most original open source LLMs were based on the llama architecture. And look where we are now OSS modles are very close to frontier.

It may not have benefitted Meta but it commoditizatised LLMs.

reply
Hell, most of us are still using llama.cpp for inference in some form
reply
The llama weights were leaked. It open sourced itself.

You are right though. Meta could have been in lockstep releasing ChatGPT features into some chat bot on Facebook.com but instead it seemed like their FAIR arm was hell bent on commoditising this stuff by publishing their research models before the Chinese companies took the lead in that.

It’s hard for me to be mad at FAIR even though I general disagree with the outcomes that Meta produce for their users.

reply
deleted
reply