points
It sounds like "we know the LLM understood its actions... because it understood its actions when we trained it", which is circular-logic.