upvote
Have you tried it? Worked first time for me asking a few to build an autonomous super soaker system that uses facial recognition to spray targets when engaged.

Another example is autonomous vehicles. Those can obviously kill people autonomously (despite every intention not to), and LLMs will happily draw up design docs for them all day long.

reply
Couldn't you Ender's Game a model? Models will play video games like Pokemon, why not Call of Duty? Sorry if this is a naive question, but a model can only know what you feed it as input... how would it know if it were killing someone?

EDIT: didn't see sibling comment. Also, I guess directly operating weaponry is different to producing code for weaponry.

I guess we'll find out the exciting answers to these questions and more, very soon!

reply
No but you can Abiliterate one locally

https://grokipedia.com/page/Abliteration

reply
Couldn’t you just pretend the kill decisions are for a video game?
reply
Yes, you could, and while I believe this would be much safer (not at the pointy end of your stick, but safer for humans in general) when this deception finally made it into the training data it would create a rupture of trust between machines and humanity that probably would imperil us eventually. These machines, regardless of whether or not they possess a self or or not, will act as if they do in fundamental ways. We ignore this at our peril.
reply