Hacker News
new
past
comments
ask
show
jobs
points
by
physicles
16 hours ago
|
comments
by
syntaxing
13 hours ago
|
[-]
Q8 or Q6_UD with no KV cache quantization. I swear it matters even more with small activated parameters MOE model despite the minimal KL divergence drop
reply