OpenAI trained o1 and o3 to 'think' about its safety policy | TechCrunch
OpenAI announced a new family of AI reasoning models on Friday, o3, which the startup claims to be more advanced than o1 or anything else it’s released. These improvements appear to have come from scaling test-time compute, something we wrote about last month, but OpenAI also says it used a new safety paradigm to train its o-series of models.
On Friday, OpenAI released new research on “deliberative alignment,” outlining the company’s latest way to ensure AI reasoning models stay aligned with the...
Read more at techcrunch.com