It remains to be seen whether Anthropic's models are still usable now.
I know just how much of a clusterfuck their "CBRN filter" is, so I'm dreading the worst.
I'd argue that black hats will find a way to get uncensored models and use them to write malware either way, and that further restricting generally available LLMs for cybersec usage would end up hurting white hats and programmers pentesting their own code way more (which would once again help the black hats, as they would have an advantage at finding unpatched exploits).