welcome
TechCrunch

TechCrunch

Technology

Technology

OpenAI trained o1 and o3 to 'think' about its safety policy | TechCrunch

TechCrunch
Summary
Nutrition label

80% Informative

OpenAI announced a new family of AI reasoning models on Friday , o3, which the startup claims to be more advanced than o1 or anything else it’s released.

The startup used this method to make o1 and o3 “think” about its safety policy during inference, the phase after a user presses enter on their prompt.

This method improved the rate at which o1 answered “unsafe” questions at least ones deemed unsafe by OpenAI .

OpenAI says its o-series of AI models have improved safety responses.

The o3 model is set to rollout sometime in 2025 .

Deliberative alignment was developed without using human-written answers or chain-of-thoughts.

The company says it used synthetic data to train the models on these examples.

VR Score

78

Informative language

74

Neutral language

80

Article tone

semi-formal

Language

English

Language complexity

56

Offensive language

not offensive

Hate speech

not hateful

Attention-grabbing headline

not detected

Known propaganda techniques

not detected

Time-value

long-living

Affiliate links

no affiliate links