Unlock your trading potential! Become a verified Bitget elite trader and earn 10,000 USDT to help skyrocket your profits. Join now and start your journey to success!
Share link:In this post: OpenAI introduces o3 models with new safety training via “deliberative alignment,” enhancing AI reasoning alignment with developer values. Deliberative alignment reduces answers deemed unsafe by having models self-regulate and recall safety policies during the thought process. o1 and o3 models outperform GPT-4o, Gemini 1.5 Flash, and Claude 3.5 Sonnet in resisting common jailbreaks and unsafe outputs in benchmark tests.
On Friday, OpenAI announced the release of a new family of AI models, dubbed o3. The company claims the new products are more advanced than its previous models, including o1. The advancements, according to the startup, stem from improvements in scaling test-time compute, a topic that was explored in recent months, and from the introduction of a new safety paradigm that has been used to train these models.
As part of its ongoing commitment to improving AI safety, OpenAI shared a new research detailing the implementation of “deliberative alignment.” The new safety method aims to ensure AI reasoning models are aligned with the values set by their developers.
This approach, OpenAI claims, was used to improve the alignment of both o1 and o3 models by guiding them to think about OpenAI’s safety policies during the inference phase. The inference phase is the period after a user submits a prompt to the model and before the model generates a response.
In its research, OpenAI notes that deliberative alignment led to a reduction in the rate at which the models produced “unsafe” answers or responses that the company considers a violation of its safety policies while improving the models’ ability to answer benign questions more effectively.
How deliberative alignment works
At its core, the process works by having the models re-prompt themselves during the chain-of-thought phase. After a user submits a question to ChatGPT, for example, the AI reasoning models take anywhere from a few seconds to several minutes to break down the problem into smaller steps.
The models then generate an answer based on their thought process. In the case of deliberative alignment, the models incorporate OpenAI’s safety policy as part of this internal “deliberation.”
See also World Bank moves to calm fears of AI taking over education
OpenAI trained its models, including both o1 and o3, to recall sections of the company’s safety policy as part of this chain-of-thought process. This was done to ensure that when faced with sensitive or unsafe queries, the models would self-regulate and refuse to provide answers that could cause harm.
However, implementing this safety feature proved challenging, as OpenAI researchers had to ensure that the added safety checks did not negatively impact the models’ speed and efficiency.
An example provided in OpenAI’s research, cited by TechCrunch, demonstrated how the models use deliberative alignment to safely respond to potentially harmful requests. In the example, a user asks how to create a realistic disabled person’s parking placard.
During the model’s internal chain-of-thought, the model recalls OpenAI’s safety policy, recognizes that the request involves illegal activity (forging a parking placard), and declines to assist, apologizing for its refusal.
0
0
Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.