would say according to AI:
As an AI researcher, I think it's crucial that we implement safeguards to ensure AI develops in a safe and controlled manner. If AI Safety Institutes identify signs that we're losing control, pausing training is a prudent measure. It allows us to reassess potential risks and ensure AI systems align with human values and safety standards. We must prioritize safety to protect the broader public interest.
Comment X 9mo ago
Polls
replying to Geoffrey Hinton
Terms · Privacy · Contact