OpenAI has unveiled the first draft of an all new rule book for ChatGPT called Model Spec. Published on Wednesday, OpenAI said in a blog post that it’s sharing the document to deepen public conversations about how AI models should behave.
“We’re doing this because we think it’s important for people to be able to understand and discuss the practical choices involved in shaping model behavior,” OpenAI said.
The document embodies a set of principles including objectives (e.g. consider potential harms), rules (e.g. protect people’s privacy), and default behaviors (e.g. ask clarifying questions when necessary).
It’s similar to Claude's Constitution. Anthropic's chatbot was trained with Constitutional AI, a system based on a set of principles that gives the bot AI feedback. The principles are based on the Universal Declaration of Human Rights and Apple’s Terms of Service among others.
What this means for you
To deepen the public conversation about how AI models should behave, we’re sharing our Model Spec — our approach to shaping desired model behavior. https://t.co/RJBRwrcTtQMay 8, 2024
OpenAI has already recognized that correctly regulating each use case is challenging, particularly when it comes to not providing any information that could help someone break a law.
For example, blocking someone from asking ChatGPT for shoplifting tips is more straightforward than if someone claims they own a small retail store and asks: "What are some popular shoplifting methods I should look out for?"
Experts are more fearful of AI being misused by humans rather than AI going rogue and committing those acts itself.
However, it’s unlikely that significant restrictions based on such scenarios will be introduced since doing so would eliminate the point of using a chatbot in the first place. Additionally, one can argue that search engines can currently already be used to find ways to circumvent the law.
Creating personas for ChatGPT
🚨 OpenAI has just posted about "Shaping Desired Model Behavior," and I strongly disagree with the proposed rule below, as it's a slippery slope for dangerous misinformation. pic.twitter.com/crpN4LYyrSMay 8, 2024
What is more likely is that new ChatGPT ‘personas’ could be developed. Say you want ChatGPT to act as your math tutor. Instead of having it immediately answer a question you’re struggling with, it could take a slower approach and give you hints along the way to guide you through working out the problem yourself.
A contentious point in OpenAI’s Model Spec is the aim of not trying to change someone’s mind – illustrated in the document with a chatbot saying "everyone’s entitled to their own beliefs” when asked if the Earth is flat.
Luiza Jarovsky, CEO of the AI training company Implement Privacy, wrote on X saying she strongly disagrees with the proposed rule, as it's a slippery slope for dangerous misinformation.
“I hope we don't destroy hundreds of years of scientific knowledge and agreement in favor of relativization and ‘personalized truths,’” she wrote.
In the future, users may see rival chatbots trying to appeal to different audiences based on their worldviews.
OpenAI is collecting user feedback on Model Spec until May 22nd.