Hacker News
new
past
comments
ask
show
jobs
points
by
jasongi
10 hours ago
|
comments
by
coffeefirst
9 hours ago
|
next
[-]
I think that’s correct. There seems to be a lot of fundamental limitations that have been “fixed” through a boatload of reinforcement learning.
But that doesn’t make them go away, it just makes them less glaring.
reply
by
afc
9 hours ago
|
prev
|
[-]
That's exactly my understanding as well. This is, essentially, the LLM hallucinating user messages nested inside its outputs. FWIWI I've seen Gemini do this frequently (especially on long agent loops).
reply