Hacker News
new
past
comments
ask
show
jobs
points
by
exabrial
23 hours ago
|
comments
by
garethsprice
11 hours ago
|
next
[-]
Something like this? (Llama 3.1-8B etched into custom silicon delivering 16,000 tok/s, doesn't use much PCIe bandwidth):
-
https://taalas.com/the-path-to-ubiquitous-ai/
-
https://chatjimmy.ai/
reply
by
exabrial
9 hours ago
|
parent
|
[-]
Wowsa that’s amazing! Exactly what I was imagining. To do that with 2500 watts is incredible.
reply
by
23 hours ago
|
prev
|
[-]
deleted
reply