ChatGPT publisher forms team of experts to assess most serious threats associated with advanced AI models.

At the end of October 2023, OpenAI announced the creation of a team of experts, named “Preparedness”, dedicated to fighting the “catastrophic” risks associated with AI. “Frontier AI models, which will exceed the capabilities currently present in the most advanced existing models, have the potential to benefit all of humanity. But they also pose increasingly severe risks,” said the ChatGPT publisher.

The team, led by Aleksander Madry, an AI expert from MIT, will focus on three main questions:

  • When purposefully misused, just how dangerous are the frontier AI systems we have today and those coming in the future?”
  • If frontier AI model weights were stolen, what exactly could a malicious actor do?”
  • How can a framework that monitors, evaluates, predicts, and protects against the dangerous capabilities of frontier AI systems be built?”

Specifically, Preparedness will study several risks associated with AI models:

  • individualized persuasion (or matching the content of a message to what the recipient wants to hear);
  • overall cybersecurity;
  • autonomous replication and adaptation (AI changing itself on its own)
  • extinction-level threats like chemical, biological, radiological, and nuclear attacks

OpenAI also launched the “AI Preparedness Challenge”. The company is offering up to $25,000 to the top 10 submissions of probable, but potentially catastrophic misuses of OpenAI.

Stay tuned in real time
Subscribe to
the newsletter
By providing your email address you agree to receive the Incyber newsletter and you have read our privacy policy. You can unsubscribe at any time by clicking on the unsubscribe link in all our emails.
Stay tuned in real time
Subscribe to
the newsletter
By providing your email address you agree to receive the Incyber newsletter and you have read our privacy policy. You can unsubscribe at any time by clicking on the unsubscribe link in all our emails.