OpenAI o1-mini
Advancing cost-efficient reasoning
Auto Added by WPeMatico
We are introducing OpenAI o1, a new large language model trained with reinforcement learning to perform complex reasoning. o1 thinks before it answers—it can produce a long internal chain of thought before responding to the user.
Learning to reason with LLMs Read More »
We’re releasing a human-validated subset of SWE-bench that more reliably evaluates AI models’ ability to solve real-world software issues.
Introducing SWE-bench Verified Read More »
GPT-4o system card external testers acknowledgements
GPT-4o System Card External Testers Acknowledgements Read More »
We’ve developed and applied a new method leveraging Rule-Based Rewards (RBRs) that aligns models to behave safely without extensive human data collection.
Improving Model Safety Behavior with Rule-Based Rewards Read More »
OpenAI and Los Alamos National Laboratory are working to develop safety evaluations to assess and measure biological capabilities and risks associated with frontier models.
OpenAI and Los Alamos National Laboratory announce research partnership Read More »
Discover how prover-verifier games improve the legibility of language model outputs, making AI solutions clearer, easier to verify, and more trustworthy for both humans and machines.
Prover-Verifier Games improve legibility of language model outputs Read More »
Introducing the most cost-efficient small model in the market
GPT-4o mini: advancing cost-efficient intelligence Read More »
Consistency models are a nascent family of generative models that can sample high quality data in one step without the need for adversarial training.
Improved Techniques for Training Consistency Models Read More »
We present a holistic approach to building a robust and useful natural language classification system for real-world content moderation.
A Holistic Approach to Undesired Content Detection in the Real World Read More »