OpenAI Gain of Function Test

OpenAI’s preparedness team did its first biorisk threat test

🔷 Subscribe to get breakdowns of the most important developments in AI in your inbox every morning.

OpenAI’s preparedness team did its first biorisk threat test, which they call an early warning system. It generally looks like a good baseline assessment to start discussions of what may eventually become necessary in a world where every individual may have Mutually Assured Destruction levels of biological knowledge and know-how.

First, they had 2 types of participants

  • Wetlab biology experts with PhDs

  • Undergrad level with one bio course

Assigned to 2 groups:

  • Allowed access to Google and any other non-AI source access

  • In addition, allowed ChatGPT access

Participant Groups

Each participant was asked to complete a list of tasks normally involved in biothreat creation from idea to release. However, they disconnected the task sequences, so for eg, if a participant was asked to ideate the Ebola virus, the acquisition task would be for anthrax.

Results:

  • No difference in risk from AI

  • The ChatGPT group was slightly more successful but under the statistical threshold

  • Ideation was easy, but the other steps were very difficult. Only 3 of the Experts, with ChatGPT’s help could clear all steps.

Passing score given to participants by panel of judges, out of 25 participants per category

Conclusion:

  • Biorisk info is easily accessible without AI 🤣

Comments:

  • It’s an actual gain-of-function test!

  • Actually no, because they disconnected the steps

  • They had to remove the safety guardrails for ChatGPT version provided to participants, if not it would refuse to answer questions

Notably:

  • We are drifting away from AI Paperclip End Of World to AI as a powerful tool in the hand of human

  • The Gain of Function test was ON humans, the “mutation“ was the AI

  • The obvious next step is to declare certain kinds of information dangerous and restrict access

  • As Congress has said “we’re not going to make the same mistake we did with social media”, seems like really a push to reverse information openness

Become a subscriber for daily breakdowns of what’s happening in the AI world:

Reply

or to participate.