I missed it when it came out but it’s in the print edition of IEEE Spectrum of February, and has been in the online edition for some time.
Leading AI Companies Get Lousy Grades on Safety
A new report from the Future of Life Institute gave mostly Ds and Fs
Full report of the Future of Life Institute
The just-released AI Safety Index graded six leading AI companies on their risk assessment efforts and safety procedures… and the top of class was Anthropic, with an overall score of C. The other five companies—Google DeepMind, Meta, OpenAI, xAI, and Zhipu AI—received grades of D+ or lower, with Meta flat out failing.
“The purpose of this is not to shame anybody,” says Max Tegmark, an MIT physics professor and president of the Future of Life Institute, which put out the report. “It’s to provide incentives for companies to improve.” He hopes that company executives will view the index like universities view the U.S. News and World Reports rankings: They may not enjoy being graded, but if the grades are out there and getting attention, they’ll feel driven to do better next year.
I do no foresee the bad outcome of AI is a crapshoot just yet, but trying to check whether the LLM knows too much about biological and chemical warfare is a good idea. Then again, we have officially funded dubious ‘research’ into potentially highly dangerous modded viruses than then go AWOL. Eh, what a world!
Who is on the board:
- Yoshua Bengio (Professor, CS and OR, University Montréal)
- Atoosa Kasirzadeh (Assistant Professor, Philo. and AI, Carnegie Mellon University)
- David Krueger (Assistant Professor, CS and OR, University Montréal)
- Tegan Maharaj (Assistant Professor, Decision Sciences, HEC Montréal)
- Sneha Revanur (Founder and President of “Encode Justice”)
- Jessica Newman (Director the “AI Security Initiative” at Berkeley)
- Stuart Russell (Professor, CS, University of California at Berkeley)