Here's the most approachable paper that shows a real model (Claude 3 Sonnet) clearly having an internal representation of bugs in code: https://transformer-circuits.pub/2024/scaling-monosemanticit...
Read the entire section around this quote:
> Thus, we concluded that 1M/1013764 represents a broad variety of errors in code.
(Also the section after "We find three different safety-relevant code features: an unsafe code feature 1M/570621 which activates on security vulnerabilities, a code error feature 1M/1013764 which activates on bugs and exceptions")
This feature fires on actual bugs; it's not just a model pattern matching saying "what a bug hunter may say next".
PS: I know it is interesting and I don't doubt Antrophic, but for me it is so fascinating they get such a pass in science.
The lifeblood of the field is proof-of-concept pre-prints built on top of other proof-of-concept pre-prints.
You don't think a pattern matcher would fire on actual bugs?
On the flip side the idea of this being true has been a very successful indirect marketing campaign.
I don’t think we even have a coherent definition of human intelligence, let alone of non-human ones.