Hacker News
new
past
comments
ask
show
jobs
points
by
manmal
6 hours ago
|
comments
by
DeathArrow
5 hours ago
|
[-]
Yep, maybe I can open a feature request if it makes sense technically.
reply
by
zozbot234
4 hours ago
|
parent
|
[-]
Arguably it makes more sense technically to get the model support into llama.cpp, which provides many options for GPU+CPU split inference already.
reply