We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
Comment by Steven Adler
AI safety researcher; Lawfare writer
Before a new model's release, Al companies commonly (though not always) run safety tests - and release the results in a "System Card."
The idea is to see if the model has any extreme abilities (like strong cyberhacking), and then to take an appropriate level of caution.
[...] I'm impressed with the thoroughness of Anthropic's testing and disclosures here. I haven't yet vetted the efficacy of the measures, but there's a lot of detail. Many companies would say far less.
But it's a problem that tests like these are totally voluntary today. Anthropic says it is setting a positive example, which I generally believe. But that's not enough: Tests like these should be required across leading Al companies, as I've written about previously.
Unverified
source
(2025)
Polls
replying to Steven Adler