<?xml version='1.0' encoding='UTF-8'?>
<rss xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" version="2.0">
  <channel>
    <title>OpenAI Research</title>
    <link>https://openai.com/news</link>
    <description>Research posts filtered from the official OpenAI RSS feed</description>
    <docs>http://www.rssboard.org/rss-specification</docs>
    <generator>RSS Creator - RSS Filter</generator>
    <language>zh-CN</language>
    <lastBuildDate>Sun, 05 Apr 2026 13:52:12 +0000</lastBuildDate>
    <item>
      <title>Weight normalization: A simple reparameterization to accelerate training of deep neural networks</title>
      <link>https://openai.com/index/weight-normalization</link>
      <guid isPermaLink="true">https://openai.com/index/weight-normalization</guid>
      <category>Research</category>
      <pubDate>Thu, 25 Feb 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Gym Beta</title>
      <link>https://openai.com/index/openai-gym-beta</link>
      <description>We’re releasing the public beta of OpenAI Gym, a toolkit for developing and comparing reinforcement learning (RL) algorithms. It consists of a growing suite of environments (from simulated robots to Atari games), and a site for comparing and reproducing results.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-gym-beta</guid>
      <category>Research</category>
      <pubDate>Wed, 27 Apr 2016 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Generative models</title>
      <link>https://openai.com/index/generative-models</link>
      <description>This post describes four projects that share a common theme of enhancing or using generative models, a branch of unsupervised learning techniques in machine learning. In addition to describing our work, this post will tell you a bit more about generative models: what they are, why they are important, and where they might be going.</description>
      <guid isPermaLink="true">https://openai.com/index/generative-models</guid>
      <category>Research</category>
      <pubDate>Thu, 16 Jun 2016 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Infrastructure for deep learning</title>
      <link>https://openai.com/index/infrastructure-for-deep-learning</link>
      <description>Deep learning is an empirical science, and the quality of a group’s infrastructure is a multiplier on progress. Fortunately, today’s open-source ecosystem makes it possible for anyone to build great deep learning infrastructure.</description>
      <guid isPermaLink="true">https://openai.com/index/infrastructure-for-deep-learning</guid>
      <category>Research</category>
      <pubDate>Mon, 29 Aug 2016 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Transfer from simulation to real world through learning deep inverse dynamics model</title>
      <link>https://openai.com/index/transfer-from-simulation-to-real-world-through-learning-deep-inverse-dynamics-model</link>
      <guid isPermaLink="true">https://openai.com/index/transfer-from-simulation-to-real-world-through-learning-deep-inverse-dynamics-model</guid>
      <category>Research</category>
      <pubDate>Tue, 11 Oct 2016 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Extensions and limitations of the neural GPU</title>
      <link>https://openai.com/index/extensions-and-limitations-of-the-neural-gpu</link>
      <guid isPermaLink="true">https://openai.com/index/extensions-and-limitations-of-the-neural-gpu</guid>
      <category>Research</category>
      <pubDate>Wed, 02 Nov 2016 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Variational lossy autoencoder</title>
      <link>https://openai.com/index/variational-lossy-autoencoder</link>
      <guid isPermaLink="true">https://openai.com/index/variational-lossy-autoencoder</guid>
      <category>Research</category>
      <pubDate>Tue, 08 Nov 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>RL²: Fast reinforcement learning via slow reinforcement learning</title>
      <link>https://openai.com/index/rl2</link>
      <guid isPermaLink="true">https://openai.com/index/rl2</guid>
      <category>Research</category>
      <pubDate>Wed, 09 Nov 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>A connection between generative adversarial networks, inverse reinforcement learning, and energy-based models</title>
      <link>https://openai.com/index/a-connection-between-generative-adversarial-networks-inverse-reinforcement-learning-and-energy-based-models</link>
      <guid isPermaLink="true">https://openai.com/index/a-connection-between-generative-adversarial-networks-inverse-reinforcement-learning-and-energy-based-models</guid>
      <category>Research</category>
      <pubDate>Fri, 11 Nov 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>On the quantitative analysis of decoder-based generative models</title>
      <link>https://openai.com/index/on-the-quantitative-analysis-of-decoder-based-generative-models</link>
      <guid isPermaLink="true">https://openai.com/index/on-the-quantitative-analysis-of-decoder-based-generative-models</guid>
      <category>Research</category>
      <pubDate>Mon, 14 Nov 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>#Exploration: A study of count-based exploration for deep reinforcement learning</title>
      <link>https://openai.com/index/exploration</link>
      <guid isPermaLink="true">https://openai.com/index/exploration</guid>
      <category>Research</category>
      <pubDate>Tue, 15 Nov 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Universe</title>
      <link>https://openai.com/index/universe</link>
      <description>We’re releasing Universe, a software platform for measuring and training an AI’s general intelligence across the world’s supply of games, websites and other applications.</description>
      <guid isPermaLink="true">https://openai.com/index/universe</guid>
      <category>Research</category>
      <pubDate>Mon, 05 Dec 2016 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>PixelCNN++: Improving the PixelCNN with discretized logistic mixture likelihood and other modifications</title>
      <link>https://openai.com/index/pixelcnn-plus-plus</link>
      <guid isPermaLink="true">https://openai.com/index/pixelcnn-plus-plus</guid>
      <category>Research</category>
      <pubDate>Thu, 19 Jan 2017 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Third-person imitation learning</title>
      <link>https://openai.com/index/third-person-imitation-learning</link>
      <guid isPermaLink="true">https://openai.com/index/third-person-imitation-learning</guid>
      <category>Research</category>
      <pubDate>Mon, 06 Mar 2017 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Prediction and control with temporal segment models</title>
      <link>https://openai.com/index/prediction-and-control-with-temporal-segment-models</link>
      <guid isPermaLink="true">https://openai.com/index/prediction-and-control-with-temporal-segment-models</guid>
      <category>Research</category>
      <pubDate>Sun, 12 Mar 2017 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Emergence of grounded compositional language in multi-agent populations</title>
      <link>https://openai.com/index/emergence-of-grounded-compositional-language-in-multi-agent-populations</link>
      <guid isPermaLink="true">https://openai.com/index/emergence-of-grounded-compositional-language-in-multi-agent-populations</guid>
      <category>Research</category>
      <pubDate>Wed, 15 Mar 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning to communicate</title>
      <link>https://openai.com/index/learning-to-communicate</link>
      <description>In this post we’ll outline new OpenAI research in which agents develop their own language.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-to-communicate</guid>
      <category>Research</category>
      <pubDate>Thu, 16 Mar 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>One-shot imitation learning</title>
      <link>https://openai.com/index/one-shot-imitation-learning</link>
      <guid isPermaLink="true">https://openai.com/index/one-shot-imitation-learning</guid>
      <category>Research</category>
      <pubDate>Tue, 21 Mar 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evolution strategies as a scalable alternative to reinforcement learning</title>
      <link>https://openai.com/index/evolution-strategies</link>
      <description>We’ve discovered that evolution strategies (ES), an optimization technique that’s been known for decades, rivals the performance of standard reinforcement learning (RL) techniques on modern RL benchmarks (e.g. Atari/MuJoCo), while overcoming many of RL’s inconveniences.</description>
      <guid isPermaLink="true">https://openai.com/index/evolution-strategies</guid>
      <category>Research</category>
      <pubDate>Fri, 24 Mar 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Spam detection in the physical world</title>
      <link>https://openai.com/index/spam-detection-in-the-physical-world</link>
      <description>We’ve created the world’s first Spam-detecting AI trained entirely in simulation and deployed on a physical robot.</description>
      <guid isPermaLink="true">https://openai.com/index/spam-detection-in-the-physical-world</guid>
      <category>Research</category>
      <pubDate>Sat, 01 Apr 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Unsupervised sentiment neuron</title>
      <link>https://openai.com/index/unsupervised-sentiment-neuron</link>
      <description>We’ve developed an unsupervised system which learns an excellent representation of sentiment, despite being trained only to predict the next character in the text of Amazon reviews.</description>
      <guid isPermaLink="true">https://openai.com/index/unsupervised-sentiment-neuron</guid>
      <category>Research</category>
      <pubDate>Thu, 06 Apr 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Stochastic Neural Networks for hierarchical reinforcement learning</title>
      <link>https://openai.com/index/stochastic-neural-networks-for-hierarchical-reinforcement-learning</link>
      <guid isPermaLink="true">https://openai.com/index/stochastic-neural-networks-for-hierarchical-reinforcement-learning</guid>
      <category>Research</category>
      <pubDate>Mon, 10 Apr 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Equivalence between policy gradients and soft Q-learning</title>
      <link>https://openai.com/index/equivalence-between-policy-gradients-and-soft-q-learning</link>
      <guid isPermaLink="true">https://openai.com/index/equivalence-between-policy-gradients-and-soft-q-learning</guid>
      <category>Research</category>
      <pubDate>Fri, 21 Apr 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Roboschool</title>
      <link>https://openai.com/index/roboschool</link>
      <description>We are releasing Roboschool: open-source software for robot simulation, integrated with OpenAI Gym.</description>
      <guid isPermaLink="true">https://openai.com/index/roboschool</guid>
      <category>Research</category>
      <pubDate>Mon, 15 May 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Robots that learn</title>
      <link>https://openai.com/index/robots-that-learn</link>
      <description>We’ve created a robotics system, trained entirely in simulation and deployed on a physical robot, which can learn a new task after seeing it done once.</description>
      <guid isPermaLink="true">https://openai.com/index/robots-that-learn</guid>
      <category>Research</category>
      <pubDate>Tue, 16 May 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Baselines: DQN</title>
      <link>https://openai.com/index/openai-baselines-dqn</link>
      <description>We’re open-sourcing OpenAI Baselines, our internal effort to reproduce reinforcement learning algorithms with performance on par with published results. We’ll release the algorithms over upcoming months; today’s release includes DQN and three of its variants.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-baselines-dqn</guid>
      <category>Research</category>
      <pubDate>Wed, 24 May 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>UCB exploration via Q-ensembles</title>
      <link>https://openai.com/index/ucb-exploration-via-q-ensembles</link>
      <guid isPermaLink="true">https://openai.com/index/ucb-exploration-via-q-ensembles</guid>
      <category>Research</category>
      <pubDate>Mon, 05 Jun 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning to cooperate, compete, and communicate</title>
      <link>https://openai.com/index/learning-to-cooperate-compete-and-communicate</link>
      <description>Multiagent environments where agents compete for resources are stepping stones on the path to AGI. Multiagent environments have two useful properties: first, there is a natural curriculum—the difficulty of the environment is determined by the skill of your competitors (and if you’re competing against clones of yourself, the environment exactly matches your skill level). Second, a multiagent environment has no stable equilibrium: no matter how smart an agent is, there’s always pressure to get smarter. These environments have a very different feel from traditional environments, and it’ll take a lot more research before we become good at them.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-to-cooperate-compete-and-communicate</guid>
      <category>Research</category>
      <pubDate>Thu, 08 Jun 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Faster physics in Python</title>
      <link>https://openai.com/index/faster-physics-in-python</link>
      <description>We’re open-sourcing a high-performance Python library for robotic simulation using the MuJoCo engine, developed over our past year of robotics research.</description>
      <guid isPermaLink="true">https://openai.com/index/faster-physics-in-python</guid>
      <category>Research</category>
      <pubDate>Wed, 28 Jun 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Teacher–student curriculum learning</title>
      <link>https://openai.com/index/teacher-student-curriculum-learning</link>
      <guid isPermaLink="true">https://openai.com/index/teacher-student-curriculum-learning</guid>
      <category>Research</category>
      <pubDate>Sat, 01 Jul 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Hindsight Experience Replay</title>
      <link>https://openai.com/index/hindsight-experience-replay</link>
      <guid isPermaLink="true">https://openai.com/index/hindsight-experience-replay</guid>
      <category>Research</category>
      <pubDate>Wed, 05 Jul 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Robust adversarial inputs</title>
      <link>https://openai.com/index/robust-adversarial-inputs</link>
      <description>We’ve created images that reliably fool neural network classifiers when viewed from varied scales and perspectives. This challenges a claim from last week that self-driving cars would be hard to trick maliciously since they capture images from multiple scales, angles, perspectives, and the like.</description>
      <guid isPermaLink="true">https://openai.com/index/robust-adversarial-inputs</guid>
      <category>Research</category>
      <pubDate>Mon, 17 Jul 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Proximal Policy Optimization</title>
      <link>https://openai.com/index/openai-baselines-ppo</link>
      <description>We’re releasing a new class of reinforcement learning algorithms, Proximal Policy Optimization (PPO), which perform comparably or better than state-of-the-art approaches while being much simpler to implement and tune. PPO has become the default reinforcement learning algorithm at OpenAI because of its ease of use and good performance.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-baselines-ppo</guid>
      <category>Research</category>
      <pubDate>Thu, 20 Jul 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Better exploration with parameter noise</title>
      <link>https://openai.com/index/better-exploration-with-parameter-noise</link>
      <description>We’ve found that adding adaptive noise to the parameters of reinforcement learning algorithms frequently boosts performance. This exploration method is simple to implement and very rarely decreases performance, so it’s worth trying on any problem.</description>
      <guid isPermaLink="true">https://openai.com/index/better-exploration-with-parameter-noise</guid>
      <category>Research</category>
      <pubDate>Thu, 27 Jul 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Gathering human feedback</title>
      <link>https://openai.com/index/gathering-human-feedback</link>
      <description>RL-Teacher is an open-source implementation of our interface to train AIs via occasional human feedback rather than hand-crafted reward functions. The underlying technique was developed as a step towards safe AI systems, but also applies to reinforcement learning problems with rewards that are hard to specify.</description>
      <guid isPermaLink="true">https://openai.com/index/gathering-human-feedback</guid>
      <category>Research</category>
      <pubDate>Thu, 03 Aug 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Dota 2</title>
      <link>https://openai.com/index/dota-2</link>
      <description>We’ve created a bot which beats the world’s top professionals at 1v1 matches of Dota 2 under standard tournament rules. The bot learned the game from scratch by self-play, and does not use imitation learning or tree search. This is a step towards building AI systems which accomplish well-defined goals in messy, complicated situations involving real humans.</description>
      <guid isPermaLink="true">https://openai.com/index/dota-2</guid>
      <category>Research</category>
      <pubDate>Fri, 11 Aug 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>More on Dota 2</title>
      <link>https://openai.com/index/more-on-dota-2</link>
      <description>Our Dota 2 result shows that self-play can catapult the performance of machine learning systems from far below human level to superhuman, given sufficient compute. In the span of a month, our system went from barely matching a high-ranked player to beating the top pros and has continued to improve since then. Supervised deep learning systems can only be as good as their training datasets, but in self-play systems, the available data improves automatically as the agent gets better.</description>
      <guid isPermaLink="true">https://openai.com/index/more-on-dota-2</guid>
      <category>Research</category>
      <pubDate>Wed, 16 Aug 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Baselines: ACKTR &amp; A2C</title>
      <link>https://openai.com/index/openai-baselines-acktr-a2c</link>
      <description>We’re releasing two new OpenAI Baselines implementations: ACKTR and A2C. A2C is a synchronous, deterministic variant of Asynchronous Advantage Actor Critic (A3C) which we’ve found gives equal performance. ACKTR is a more sample-efficient reinforcement learning algorithm than TRPO and A2C, and requires only slightly more computation than A2C per update.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-baselines-acktr-a2c</guid>
      <category>Research</category>
      <pubDate>Fri, 18 Aug 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning with opponent-learning awareness</title>
      <link>https://openai.com/index/learning-with-opponent-learning-awareness</link>
      <guid isPermaLink="true">https://openai.com/index/learning-with-opponent-learning-awareness</guid>
      <category>Research</category>
      <pubDate>Wed, 13 Sep 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning to model other minds</title>
      <link>https://openai.com/index/learning-to-model-other-minds</link>
      <description>We’re releasing an algorithm which accounts for the fact that other agents are learning too, and discovers self-interested yet collaborative strategies like tit-for-tat in the iterated prisoner’s dilemma. This algorithm, Learning with Opponent-Learning Awareness (LOLA), is a small step towards agents that model other minds.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-to-model-other-minds</guid>
      <category>Research</category>
      <pubDate>Thu, 14 Sep 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Nonlinear computation in deep linear networks</title>
      <link>https://openai.com/index/nonlinear-computation-in-deep-linear-networks</link>
      <guid isPermaLink="true">https://openai.com/index/nonlinear-computation-in-deep-linear-networks</guid>
      <category>Research</category>
      <pubDate>Fri, 29 Sep 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Competitive self-play</title>
      <link>https://openai.com/index/competitive-self-play</link>
      <description>We’ve found that self-play allows simulated AIs to discover physical skills like tackling, ducking, faking, kicking, catching, and diving for the ball, without explicitly designing an environment with these skills in mind. Self-play ensures that the environment is always the right difficulty for an AI to improve. Taken alongside our Dota 2 self-play results, we have increasing confidence that self-play will be a core part of powerful AI systems in the future.</description>
      <guid isPermaLink="true">https://openai.com/index/competitive-self-play</guid>
      <category>Research</category>
      <pubDate>Wed, 11 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Meta-learning for wrestling</title>
      <link>https://openai.com/index/meta-learning-for-wrestling</link>
      <description>We show that for the task of simulated robot wrestling, a meta-learning agent can learn to quickly defeat a stronger non-meta-learning agent, and also show that the meta-learning agent can adapt to physical malfunction.</description>
      <guid isPermaLink="true">https://openai.com/index/meta-learning-for-wrestling</guid>
      <category>Research</category>
      <pubDate>Wed, 11 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Domain randomization and generative models for robotic grasping</title>
      <link>https://openai.com/index/domain-randomization-and-generative-models-for-robotic-grasping</link>
      <guid isPermaLink="true">https://openai.com/index/domain-randomization-and-generative-models-for-robotic-grasping</guid>
      <category>Research</category>
      <pubDate>Tue, 17 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Sim-to-real transfer of robotic control with dynamics randomization</title>
      <link>https://openai.com/index/sim-to-real-transfer-of-robotic-control-with-dynamics-randomization</link>
      <guid isPermaLink="true">https://openai.com/index/sim-to-real-transfer-of-robotic-control-with-dynamics-randomization</guid>
      <category>Research</category>
      <pubDate>Wed, 18 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Asymmetric actor critic for image-based robot learning</title>
      <link>https://openai.com/index/asymmetric-actor-critic-for-image-based-robot-learning</link>
      <guid isPermaLink="true">https://openai.com/index/asymmetric-actor-critic-for-image-based-robot-learning</guid>
      <category>Research</category>
      <pubDate>Wed, 18 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Generalizing from simulation</title>
      <link>https://openai.com/index/generalizing-from-simulation</link>
      <description>Our latest robotics techniques allow robot controllers, trained entirely in simulation and deployed on physical robots, to react to unplanned changes in the environment as they solve simple tasks. That is, we’ve used these techniques to build closed-loop systems rather than open-loop ones as before.</description>
      <guid isPermaLink="true">https://openai.com/index/generalizing-from-simulation</guid>
      <category>Research</category>
      <pubDate>Thu, 19 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning a hierarchy</title>
      <link>https://openai.com/index/learning-a-hierarchy</link>
      <description>We’ve developed a hierarchical reinforcement learning algorithm that learns high-level actions useful for solving a range of tasks, allowing fast solving of tasks requiring thousands of timesteps. Our algorithm, when applied to a set of navigation problems, discovers a set of high-level actions for walking and crawling in different directions, which enables the agent to master new navigation tasks quickly.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-a-hierarchy</guid>
      <category>Research</category>
      <pubDate>Thu, 26 Oct 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Interpretable and pedagogical examples</title>
      <link>https://openai.com/index/interpretable-and-pedagogical-examples</link>
      <guid isPermaLink="true">https://openai.com/index/interpretable-and-pedagogical-examples</guid>
      <category>Research</category>
      <pubDate>Thu, 02 Nov 2017 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning sparse neural networks through L₀ regularization</title>
      <link>https://openai.com/index/learning-sparse-neural-networks-through-l0-regularization</link>
      <guid isPermaLink="true">https://openai.com/index/learning-sparse-neural-networks-through-l0-regularization</guid>
      <category>Research</category>
      <pubDate>Mon, 04 Dec 2017 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Block-sparse GPU kernels</title>
      <link>https://openai.com/index/block-sparse-gpu-kernels</link>
      <description>We’re releasing highly-optimized GPU kernels for an underexplored class of neural network architectures: networks with block-sparse weights. Depending on the chosen sparsity, these kernels can run orders of magnitude faster than cuBLAS or cuSPARSE. We’ve used them to attain state-of-the-art results in text sentiment analysis and generative modeling of text and images.</description>
      <guid isPermaLink="true">https://openai.com/index/block-sparse-gpu-kernels</guid>
      <category>Research</category>
      <pubDate>Wed, 06 Dec 2017 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Scaling Kubernetes to 2,500 nodes</title>
      <link>https://openai.com/index/scaling-kubernetes-to-2500-nodes</link>
      <guid isPermaLink="true">https://openai.com/index/scaling-kubernetes-to-2500-nodes</guid>
      <category>Research</category>
      <pubDate>Thu, 18 Jan 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Requests for Research 2.0</title>
      <link>https://openai.com/index/requests-for-research-2</link>
      <description>We’re releasing a new batch of seven unsolved problems which have come up in the course of our research at OpenAI.</description>
      <guid isPermaLink="true">https://openai.com/index/requests-for-research-2</guid>
      <category>Research</category>
      <pubDate>Wed, 31 Jan 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Discovering types for entity disambiguation</title>
      <link>https://openai.com/index/discovering-types-for-entity-disambiguation</link>
      <description>We’ve built a system for automatically figuring out which object is meant by a word by having a neural network decide if the word belongs to each of about 100 automatically-discovered “types” (non-exclusive categories).</description>
      <guid isPermaLink="true">https://openai.com/index/discovering-types-for-entity-disambiguation</guid>
      <category>Research</category>
      <pubDate>Wed, 07 Feb 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Interpretable machine learning through teaching</title>
      <link>https://openai.com/index/interpretable-machine-learning-through-teaching</link>
      <description>We’ve designed a method that encourages AIs to teach each other with examples that also make sense to humans. Our approach automatically selects the most informative examples to teach a concept—for instance, the best images to describe the concept of dogs—and experimentally we found our approach to be effective at teaching both AIs</description>
      <guid isPermaLink="true">https://openai.com/index/interpretable-machine-learning-through-teaching</guid>
      <category>Research</category>
      <pubDate>Thu, 15 Feb 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Ingredients for robotics research</title>
      <link>https://openai.com/index/ingredients-for-robotics-research</link>
      <description>We’re releasing eight simulated robotics environments and a Baselines implementation of Hindsight Experience Replay, all developed for our research over the past year. We’ve used these environments to train models which work on physical robots. We’re also releasing a set of requests for robotics research.</description>
      <guid isPermaLink="true">https://openai.com/index/ingredients-for-robotics-research</guid>
      <category>Research</category>
      <pubDate>Mon, 26 Feb 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Multi-Goal Reinforcement Learning: Challenging robotics environments and request for research</title>
      <link>https://openai.com/index/multi-goal-reinforcement-learning</link>
      <guid isPermaLink="true">https://openai.com/index/multi-goal-reinforcement-learning</guid>
      <category>Research</category>
      <pubDate>Mon, 26 Feb 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Some considerations on learning to explore via meta-reinforcement learning</title>
      <link>https://openai.com/index/some-considerations-on-learning-to-explore-via-meta-reinforcement-learning</link>
      <guid isPermaLink="true">https://openai.com/index/some-considerations-on-learning-to-explore-via-meta-reinforcement-learning</guid>
      <category>Research</category>
      <pubDate>Sat, 03 Mar 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Reptile: A scalable meta-learning algorithm</title>
      <link>https://openai.com/index/reptile</link>
      <description>We’ve developed a simple meta-learning algorithm called Reptile which works by repeatedly sampling a task, performing stochastic gradient descent on it, and updating the initial parameters towards the final parameters learned on that task. Reptile is the application of the Shortest Descent algorithm to the meta-learning setting, and is mathematically similar to first-order MAML (which is a version of the well-known MAML algorithm) that only needs black-box access to an optimizer such as SGD or Adam, with similar computational efficiency and performance.</description>
      <guid isPermaLink="true">https://openai.com/index/reptile</guid>
      <category>Research</category>
      <pubDate>Wed, 07 Mar 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>On first-order meta-learning algorithms</title>
      <link>https://openai.com/index/on-first-order-meta-learning-algorithms</link>
      <guid isPermaLink="true">https://openai.com/index/on-first-order-meta-learning-algorithms</guid>
      <category>Research</category>
      <pubDate>Thu, 08 Mar 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improving GANs using optimal transport</title>
      <link>https://openai.com/index/improving-gans-using-optimal-transport</link>
      <guid isPermaLink="true">https://openai.com/index/improving-gans-using-optimal-transport</guid>
      <category>Research</category>
      <pubDate>Thu, 15 Mar 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Variance reduction for policy gradient with action-dependent factorized baselines</title>
      <link>https://openai.com/index/variance-reduction-for-policy-gradient-with-action-dependent-factorized-baselines</link>
      <guid isPermaLink="true">https://openai.com/index/variance-reduction-for-policy-gradient-with-action-dependent-factorized-baselines</guid>
      <category>Research</category>
      <pubDate>Tue, 20 Mar 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Retro Contest</title>
      <link>https://openai.com/index/retro-contest</link>
      <description>We’re launching a transfer learning contest that measures a reinforcement learning algorithm’s ability to generalize from previous experience.</description>
      <guid isPermaLink="true">https://openai.com/index/retro-contest</guid>
      <category>Research</category>
      <pubDate>Thu, 05 Apr 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Gotta Learn Fast: A new benchmark for generalization in RL</title>
      <link>https://openai.com/index/gotta-learn-fast</link>
      <guid isPermaLink="true">https://openai.com/index/gotta-learn-fast</guid>
      <category>Research</category>
      <pubDate>Tue, 10 Apr 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evolved Policy Gradients</title>
      <link>https://openai.com/index/evolved-policy-gradients</link>
      <description>We’re releasing an experimental metalearning approach called Evolved Policy Gradients, a method that evolves the loss function of learning agents, which can enable fast training on novel tasks. Agents trained with EPG can succeed at basic tasks at test time that were outside their training regime, like learning to navigate to an object on a different side of the room from where it was placed during training.</description>
      <guid isPermaLink="true">https://openai.com/index/evolved-policy-gradients</guid>
      <category>Research</category>
      <pubDate>Wed, 18 Apr 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>AI and compute</title>
      <link>https://openai.com/index/ai-and-compute</link>
      <description>We’re releasing an analysis showing that since 2012, the amount of compute used in the largest AI training runs has been increasing exponentially with a 3.4-month doubling time (by comparison, Moore’s Law had a 2-year doubling period)[^footnote-correction]. Since 2012, this metric has grown by more than 300,000x (a 2-year doubling period would yield only a 7x increase). Improvements in compute have been a key component of AI progress, so as long as this trend continues, it’s worth preparing for the implications of systems far outside today’s capabilities.</description>
      <guid isPermaLink="true">https://openai.com/index/ai-and-compute</guid>
      <category>Research</category>
      <pubDate>Wed, 16 May 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Gym Retro</title>
      <link>https://openai.com/index/gym-retro</link>
      <description>We’re releasing the full version of Gym Retro, a platform for reinforcement learning research on games. This brings our publicly-released game count from around 70 Atari games and 30 Sega games to over 1,000 games across a variety of backing emulators. We’re also releasing the tool we use to add new games to the platform.</description>
      <guid isPermaLink="true">https://openai.com/index/gym-retro</guid>
      <category>Research</category>
      <pubDate>Fri, 25 May 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GamePad: A learning environment for theorem proving</title>
      <link>https://openai.com/index/gamepad</link>
      <guid isPermaLink="true">https://openai.com/index/gamepad</guid>
      <category>Research</category>
      <pubDate>Sat, 02 Jun 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning policy representations in multiagent systems</title>
      <link>https://openai.com/index/learning-policy-representations-in-multiagent-systems</link>
      <guid isPermaLink="true">https://openai.com/index/learning-policy-representations-in-multiagent-systems</guid>
      <category>Research</category>
      <pubDate>Sun, 17 Jun 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Retro Contest: Results</title>
      <link>https://openai.com/index/retro-contest-results</link>
      <description>The first run of our Retro Contest—exploring the development of algorithms that can generalize from previous experience—is now complete.</description>
      <guid isPermaLink="true">https://openai.com/index/retro-contest-results</guid>
      <category>Research</category>
      <pubDate>Fri, 22 Jun 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Five</title>
      <link>https://openai.com/index/openai-five</link>
      <description>Our team of five neural networks, OpenAI Five, has started to defeat amateur human teams at Dota 2.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-five</guid>
      <category>Research</category>
      <pubDate>Mon, 25 Jun 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning Montezuma’s Revenge from a single demonstration</title>
      <link>https://openai.com/index/learning-montezumas-revenge-from-a-single-demonstration</link>
      <description>We’ve trained an agent to achieve a high score of 74,500 on Montezuma’s Revenge from a single human demonstration, better than any previously published result. Our algorithm is simple: the agent plays a sequence of games starting from carefully chosen states from the demonstration, and learns from them by optimizing the game score using PPO, the same reinforcement learning algorithm that underpins OpenAI Five.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-montezumas-revenge-from-a-single-demonstration</guid>
      <category>Research</category>
      <pubDate>Wed, 04 Jul 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Glow: Better reversible generative models</title>
      <link>https://openai.com/index/glow</link>
      <description>We introduce Glow, a reversible generative model which uses invertible 1x1 convolutions. It extends previous work on reversible generative models and simplifies the architecture. Our model can generate realistic high resolution images, supports efficient sampling, and discovers features that can be used to manipulate attributes of data. We’re releasing code for the model and an online visualization tool so people can explore and build on these results.</description>
      <guid isPermaLink="true">https://openai.com/index/glow</guid>
      <category>Research</category>
      <pubDate>Mon, 09 Jul 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Variational option discovery algorithms</title>
      <link>https://openai.com/index/variational-option-discovery-algorithms</link>
      <guid isPermaLink="true">https://openai.com/index/variational-option-discovery-algorithms</guid>
      <category>Research</category>
      <pubDate>Thu, 26 Jul 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning dexterity</title>
      <link>https://openai.com/index/learning-dexterity</link>
      <description>We’ve trained a human-like robot hand to manipulate physical objects with unprecedented dexterity.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-dexterity</guid>
      <category>Research</category>
      <pubDate>Mon, 30 Jul 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Five Benchmark: Results</title>
      <link>https://openai.com/index/openai-five-benchmark-results</link>
      <description>Yesterday, OpenAI Five won a best-of-three against a team of 99.95th percentile Dota players: Blitz, Cap, Fogged, Merlini, and MoonMeander—four of whom have played Dota professionally—in front of a live audience and 100,000 concurrent livestream viewers.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-five-benchmark-results</guid>
      <category>Research</category>
      <pubDate>Mon, 06 Aug 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Large-scale study of curiosity-driven learning</title>
      <link>https://openai.com/index/large-scale-study-of-curiosity-driven-learning</link>
      <guid isPermaLink="true">https://openai.com/index/large-scale-study-of-curiosity-driven-learning</guid>
      <category>Research</category>
      <pubDate>Mon, 13 Aug 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>The International 2018: Results</title>
      <link>https://openai.com/index/the-international-2018-results</link>
      <description>OpenAI Five lost two games against top Dota 2 players at The International in Vancouver this week, maintaining a good chance of winning for the first 20–35 minutes of both games.</description>
      <guid isPermaLink="true">https://openai.com/index/the-international-2018-results</guid>
      <category>Research</category>
      <pubDate>Thu, 23 Aug 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>FFJORD: Free-form continuous dynamics for scalable reversible generative models</title>
      <link>https://openai.com/index/ffjord</link>
      <guid isPermaLink="true">https://openai.com/index/ffjord</guid>
      <category>Research</category>
      <pubDate>Tue, 02 Oct 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Reinforcement learning with prediction-based rewards</title>
      <link>https://openai.com/index/reinforcement-learning-with-prediction-based-rewards</link>
      <description>We’ve developed Random Network Distillation (RND), a prediction-based method for encouraging reinforcement learning agents to explore their environments through curiosity, which for the first time exceeds average human performance on Montezuma’s Revenge.</description>
      <guid isPermaLink="true">https://openai.com/index/reinforcement-learning-with-prediction-based-rewards</guid>
      <category>Research</category>
      <pubDate>Wed, 31 Oct 2018 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Plan online, learn offline: Efficient learning and exploration via model-based control</title>
      <link>https://openai.com/index/plan-online-learn-offline</link>
      <guid isPermaLink="true">https://openai.com/index/plan-online-learn-offline</guid>
      <category>Research</category>
      <pubDate>Mon, 05 Nov 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning concepts with energy functions</title>
      <link>https://openai.com/index/learning-concepts-with-energy-functions</link>
      <description>We’ve developed an energy-based model that can quickly learn to identify and generate instances of concepts, such as near, above, between, closest, and furthest, expressed as sets of 2d points. Our model learns these concepts after only five demonstrations. We also show cross-domain transfer: we use concepts learned in a 2d particle environment to solve tasks on a 3-dimensional physics-based robot.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-concepts-with-energy-functions</guid>
      <category>Research</category>
      <pubDate>Wed, 07 Nov 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Spinning Up in Deep RL</title>
      <link>https://openai.com/index/spinning-up-in-deep-rl</link>
      <description>We’re releasing Spinning Up in Deep RL, an educational resource designed to let anyone learn to become a skilled practitioner in deep reinforcement learning. Spinning Up consists of crystal-clear examples of RL code, educational exercises, documentation, and tutorials.</description>
      <guid isPermaLink="true">https://openai.com/index/spinning-up-in-deep-rl</guid>
      <category>Research</category>
      <pubDate>Thu, 08 Nov 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Quantifying generalization in reinforcement learning</title>
      <link>https://openai.com/index/quantifying-generalization-in-reinforcement-learning</link>
      <description>We’re releasing CoinRun, a training environment which provides a metric for an agent’s ability to transfer its experience to novel situations and has already helped clarify a longstanding puzzle in reinforcement learning. CoinRun strikes a desirable balance in complexity: the environment is simpler than traditional platformer games like Sonic the Hedgehog but still poses a worthy generalization challenge for state of the art algorithms.</description>
      <guid isPermaLink="true">https://openai.com/index/quantifying-generalization-in-reinforcement-learning</guid>
      <category>Research</category>
      <pubDate>Thu, 06 Dec 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>How AI training scales</title>
      <link>https://openai.com/index/how-ai-training-scales</link>
      <description>We’ve discovered that the gradient noise scale, a simple statistical metric, predicts the parallelizability of neural network training on a wide range of tasks. Since complex tasks tend to have noisier gradients, increasingly large batch sizes are likely to become useful in the future, removing one potential limit to further growth of AI systems. More broadly, these results show that neural network training need not be considered a mysterious art, but can be rigorized and systematized.</description>
      <guid isPermaLink="true">https://openai.com/index/how-ai-training-scales</guid>
      <category>Research</category>
      <pubDate>Fri, 14 Dec 2018 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Computational limitations in robust classification and win-win results</title>
      <link>https://openai.com/index/computational-limitations-in-robust-classification-and-win-win-results</link>
      <guid isPermaLink="true">https://openai.com/index/computational-limitations-in-robust-classification-and-win-win-results</guid>
      <category>Research</category>
      <pubDate>Mon, 04 Feb 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Better language models and their implications</title>
      <link>https://openai.com/index/better-language-models</link>
      <description>We’ve trained a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension, machine translation, question answering, and summarization—all without task-specific training.</description>
      <guid isPermaLink="true">https://openai.com/index/better-language-models</guid>
      <category>Research</category>
      <pubDate>Thu, 14 Feb 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Neural MMO: A massively multiagent game environment</title>
      <link>https://openai.com/index/neural-mmo</link>
      <description>We’re releasing a Neural MMO, a massively multiagent game environment for reinforcement learning agents. Our platform supports a large, variable number of agents within a persistent and open-ended task. The inclusion of many agents and species leads to better exploration, divergent niche formation, and greater overall competence.</description>
      <guid isPermaLink="true">https://openai.com/index/neural-mmo</guid>
      <category>Research</category>
      <pubDate>Mon, 04 Mar 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Implicit generation and generalization methods for energy-based models</title>
      <link>https://openai.com/index/energy-based-models</link>
      <description>We’ve made progress towards stable and scalable training of energy-based models (EBMs) resulting in better sample quality and generalization ability than existing models. Generation in EBMs spends more compute to continually refine its answers and doing so can generate samples competitive with GANs at low temperatures, while also having mode coverage guarantees of likelihood-based models. We hope these findings stimulate further research into this promising class of models.</description>
      <guid isPermaLink="true">https://openai.com/index/energy-based-models</guid>
      <category>Research</category>
      <pubDate>Thu, 21 Mar 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Five defeats Dota 2 world champions</title>
      <link>https://openai.com/index/openai-five-defeats-dota-2-world-champions</link>
      <description>OpenAI Five is the first AI to beat the world champions in an esports game, having won two back-to-back games versus the world champion Dota 2 team, OG, at Finals this weekend. Both OpenAI Five and DeepMind’s AlphaStar had previously beaten good pros privately but lost their live pro matches, making this also the first time an AI has beaten esports pros on livestream.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-five-defeats-dota-2-world-champions</guid>
      <category>Research</category>
      <pubDate>Mon, 15 Apr 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Generative modeling with sparse transformers</title>
      <link>https://openai.com/index/sparse-transformer</link>
      <description>We’ve developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether text, images, or sound. It uses an algorithmic improvement of the attention mechanism to extract patterns from sequences 30x longer than possible previously.</description>
      <guid isPermaLink="true">https://openai.com/index/sparse-transformer</guid>
      <category>Research</category>
      <pubDate>Tue, 23 Apr 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>MuseNet</title>
      <link>https://openai.com/index/musenet</link>
      <description>We’ve created MuseNet, a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country to Mozart to the Beatles. MuseNet was not explicitly programmed with our understanding of music, but instead discovered patterns of harmony, rhythm, and style by learning to predict the next token in hundreds of thousands of MIDI files. MuseNet uses the same general-purpose unsupervised technology as GPT-2, a large-scale transformer model trained to predict the next token in a sequence, whether audio or text.</description>
      <guid isPermaLink="true">https://openai.com/index/musenet</guid>
      <category>Research</category>
      <pubDate>Thu, 25 Apr 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-2: 6-month follow-up</title>
      <link>https://openai.com/index/gpt-2-6-month-follow-up</link>
      <description>We’re releasing the 774 million parameter GPT-2 language model after the release of our small 124M model in February, staged release of our medium 355M model in May, and subsequent research with partners and the AI community into the model’s potential for misuse and societal benefit. We’re also releasing an open-source legal agreement to make it easier for organizations to initiate model-sharing partnerships with each other, and are publishing a technical report about our experience in coordinating with the wider AI research community on publication norms.</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-2-6-month-follow-up</guid>
      <category>Research</category>
      <pubDate>Tue, 20 Aug 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Emergent tool use from multi-agent interaction</title>
      <link>https://openai.com/index/emergent-tool-use</link>
      <description>We’ve observed agents discovering progressively more complex tool use while playing a simple game of hide-and-seek. Through training in our new simulated hide-and-seek environment, agents build a series of six distinct strategies and counterstrategies, some of which we did not know our environment supported. The self-supervised emergent complexity in this simple environment further suggests that multi-agent co-adaptation may one day produce extremely complex and intelligent behavior.</description>
      <guid isPermaLink="true">https://openai.com/index/emergent-tool-use</guid>
      <category>Research</category>
      <pubDate>Tue, 17 Sep 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Solving Rubik’s Cube with a robot hand</title>
      <link>https://openai.com/index/solving-rubiks-cube</link>
      <description>We’ve trained a pair of neural networks to solve the Rubik’s Cube with a human-like robot hand. The neural networks are trained entirely in simulation, using the same reinforcement learning code as OpenAI Five paired with a new technique called Automatic Domain Randomization (ADR). The system can handle situations it never saw during training, such as being prodded by a stuffed giraffe. This shows that reinforcement learning isn’t just a tool for virtual tasks, but can solve physical-world problems requiring unprecedented dexterity.</description>
      <guid isPermaLink="true">https://openai.com/index/solving-rubiks-cube</guid>
      <category>Research</category>
      <pubDate>Tue, 15 Oct 2019 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-2: 1.5B release</title>
      <link>https://openai.com/index/gpt-2-1-5b-release</link>
      <description>As the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5B parameters) of GPT-2 along with code and model weights to facilitate detection of outputs of GPT-2 models. While there have been larger language models released since August, we’ve continued with our original staged release plan in order to provide the community with a test case of a full staged release process. We hope that this test case will be useful to developers of future powerful models, and we’re actively continuing the conversation with the AI community on responsible publication.</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-2-1-5b-release</guid>
      <category>Research</category>
      <pubDate>Tue, 05 Nov 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Procgen Benchmark</title>
      <link>https://openai.com/index/procgen-benchmark</link>
      <description>We’re releasing Procgen Benchmark, 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizable skills.</description>
      <guid isPermaLink="true">https://openai.com/index/procgen-benchmark</guid>
      <category>Research</category>
      <pubDate>Tue, 03 Dec 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Deep double descent</title>
      <link>https://openai.com/index/deep-double-descent</link>
      <description>We show that the double descent phenomenon occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or training time. This effect is often avoided through careful regularization. While this behavior appears to be fairly universal, we don’t yet fully understand why it happens, and view further study of this phenomenon as an important research direction.</description>
      <guid isPermaLink="true">https://openai.com/index/deep-double-descent</guid>
      <category>Research</category>
      <pubDate>Thu, 05 Dec 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Dota 2 with large scale deep reinforcement learning</title>
      <link>https://openai.com/index/dota-2-with-large-scale-deep-reinforcement-learning</link>
      <guid isPermaLink="true">https://openai.com/index/dota-2-with-large-scale-deep-reinforcement-learning</guid>
      <category>Research</category>
      <pubDate>Fri, 13 Dec 2019 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Scaling laws for neural language models</title>
      <link>https://openai.com/index/scaling-laws-for-neural-language-models</link>
      <guid isPermaLink="true">https://openai.com/index/scaling-laws-for-neural-language-models</guid>
      <category>Research</category>
      <pubDate>Thu, 23 Jan 2020 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI Microscope</title>
      <link>https://openai.com/index/microscope</link>
      <description>We’re introducing OpenAI Microscope, a collection of visualizations of every significant layer and neuron of eight vision “model organisms” which are often studied in interpretability. Microscope makes it easier to analyze the features that form inside these neural networks, and we hope it will help the research community as we move towards understanding these complicated systems.</description>
      <guid isPermaLink="true">https://openai.com/index/microscope</guid>
      <category>Research</category>
      <pubDate>Tue, 14 Apr 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improving verifiability in AI development</title>
      <link>https://openai.com/index/improving-verifiability</link>
      <description>We’ve contributed to a multi-stakeholder report by 58 co-authors at 30 organizations, including the Centre for the Future of Intelligence, Mila, Schwartz Reisman Institute for Technology and Society, Center for Advanced Study in the Behavioral Sciences, and Center for Security and Emerging Technologies. This report describes 10 mechanisms to improve the verifiability of claims made about AI systems. Developers can use these tools to provide evidence that AI systems are safe, secure, fair, or privacy-preserving. Users, policymakers, and civil society can use these tools to evaluate AI development processes.</description>
      <guid isPermaLink="true">https://openai.com/index/improving-verifiability</guid>
      <category>Research</category>
      <pubDate>Thu, 16 Apr 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Jukebox</title>
      <link>https://openai.com/index/jukebox</link>
      <description>We’re introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles. We’re releasing the model weights and code, along with a tool to explore the generated samples.</description>
      <guid isPermaLink="true">https://openai.com/index/jukebox</guid>
      <category>Research</category>
      <pubDate>Thu, 30 Apr 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>AI and efficiency</title>
      <link>https://openai.com/index/ai-and-efficiency</link>
      <description>We’re releasing an analysis showing that since 2012 the amount of compute needed to train a neural net to the same performance on ImageNet classification has been decreasing by a factor of 2 every 16 months. Compared to 2012, it now takes 44 times less compute to train a neural network to the level of AlexNet (by contrast, Moore’s Law would yield an 11x cost improvement over this period). Our results suggest that for AI tasks with high levels of recent investment, algorithmic progress has yielded more gains than classical hardware efficiency.</description>
      <guid isPermaLink="true">https://openai.com/index/ai-and-efficiency</guid>
      <category>Research</category>
      <pubDate>Tue, 05 May 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Language models are few-shot learners</title>
      <link>https://openai.com/index/language-models-are-few-shot-learners</link>
      <guid isPermaLink="true">https://openai.com/index/language-models-are-few-shot-learners</guid>
      <category>Research</category>
      <pubDate>Thu, 28 May 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Image GPT</title>
      <link>https://openai.com/index/image-gpt</link>
      <description>We find that, just as a large transformer model trained on language can generate coherent text, the same exact model trained on pixel sequences can generate coherent image completions and samples. By establishing a correlation between sample quality and image classification accuracy, we show that our best generative model also contains features competitive with top convolutional nets in the unsupervised setting.</description>
      <guid isPermaLink="true">https://openai.com/index/image-gpt</guid>
      <category>Research</category>
      <pubDate>Wed, 17 Jun 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Generative language modeling for automated theorem proving</title>
      <link>https://openai.com/index/generative-language-modeling-for-automated-theorem-proving</link>
      <guid isPermaLink="true">https://openai.com/index/generative-language-modeling-for-automated-theorem-proving</guid>
      <category>Research</category>
      <pubDate>Mon, 07 Sep 2020 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>DALL·E: Creating images from text</title>
      <link>https://openai.com/index/dall-e</link>
      <description>We’ve trained a neural network called DALL·E that creates images from text captions for a wide range of concepts expressible in natural language.</description>
      <guid isPermaLink="true">https://openai.com/index/dall-e</guid>
      <category>Research</category>
      <pubDate>Tue, 05 Jan 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>CLIP: Connecting text and images</title>
      <link>https://openai.com/index/clip</link>
      <description>We’re introducing a neural network called CLIP which efficiently learns visual concepts from natural language supervision. CLIP can be applied to any visual classification benchmark by simply providing the names of the visual categories to be recognized, similar to the “zero-shot” capabilities of GPT-2 and GPT-3.</description>
      <guid isPermaLink="true">https://openai.com/index/clip</guid>
      <category>Research</category>
      <pubDate>Tue, 05 Jan 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Scaling Kubernetes to 7,500 nodes</title>
      <link>https://openai.com/index/scaling-kubernetes-to-7500-nodes</link>
      <description>We’ve scaled Kubernetes clusters to 7,500 nodes, producing a scalable infrastructure for large models like GPT-3, CLIP, and DALL·E, but also for rapid small-scale iterative research such as Scaling Laws for Neural Language Models.</description>
      <guid isPermaLink="true">https://openai.com/index/scaling-kubernetes-to-7500-nodes</guid>
      <category>Research</category>
      <pubDate>Mon, 25 Jan 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Understanding the capabilities, limitations, and societal impact of large language models</title>
      <link>https://openai.com/index/understanding-the-capabilities-limitations-and-societal-impact-of-large-language-models</link>
      <guid isPermaLink="true">https://openai.com/index/understanding-the-capabilities-limitations-and-societal-impact-of-large-language-models</guid>
      <category>Research</category>
      <pubDate>Thu, 04 Feb 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Multimodal neurons in artificial neural networks</title>
      <link>https://openai.com/index/multimodal-neurons</link>
      <description>We’ve discovered neurons in CLIP that respond to the same concept whether presented literally, symbolically, or conceptually. This may explain CLIP’s accuracy in classifying surprising visual renditions of concepts, and is also an important step toward understanding the associations and biases that CLIP and similar models learn.</description>
      <guid isPermaLink="true">https://openai.com/index/multimodal-neurons</guid>
      <category>Research</category>
      <pubDate>Thu, 04 Mar 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evaluating large language models trained on code</title>
      <link>https://openai.com/index/evaluating-large-language-models-trained-on-code</link>
      <guid isPermaLink="true">https://openai.com/index/evaluating-large-language-models-trained-on-code</guid>
      <category>Research</category>
      <pubDate>Wed, 07 Jul 2021 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing Triton: Open-source GPU programming for neural networks</title>
      <link>https://openai.com/index/triton</link>
      <description>We’re releasing Triton 1.0, an open-source Python-like programming language which enables researchers with no CUDA experience to write highly efficient GPU code—most of the time on par with what an expert would be able to produce.</description>
      <guid isPermaLink="true">https://openai.com/index/triton</guid>
      <category>Research</category>
      <pubDate>Wed, 28 Jul 2021 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>TruthfulQA: Measuring how models mimic human falsehoods</title>
      <link>https://openai.com/index/truthfulqa</link>
      <guid isPermaLink="true">https://openai.com/index/truthfulqa</guid>
      <category>Research</category>
      <pubDate>Wed, 08 Sep 2021 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Solving math word problems</title>
      <link>https://openai.com/index/solving-math-word-problems</link>
      <description>We’ve trained a system that solves grade school math problems with nearly twice the accuracy of a fine-tuned GPT-3 model. It solves about 90% as many problems as real kids: a small sample of 9-12 year olds scored 60% on a test from our dataset, while our system scored 55% on those same problems.</description>
      <guid isPermaLink="true">https://openai.com/index/solving-math-word-problems</guid>
      <category>Research</category>
      <pubDate>Fri, 29 Oct 2021 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>WebGPT: Improving the factual accuracy of language models through web browsing</title>
      <link>https://openai.com/index/webgpt</link>
      <description>We’ve fine-tuned GPT-3 to more accurately answer open-ended questions using a text-based web browser.</description>
      <guid isPermaLink="true">https://openai.com/index/webgpt</guid>
      <category>Research</category>
      <pubDate>Thu, 16 Dec 2021 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Text and code embeddings by contrastive pre-training</title>
      <link>https://openai.com/index/text-and-code-embeddings-by-contrastive-pre-training</link>
      <guid isPermaLink="true">https://openai.com/index/text-and-code-embeddings-by-contrastive-pre-training</guid>
      <category>Research</category>
      <pubDate>Mon, 24 Jan 2022 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Solving (some) formal math olympiad problems</title>
      <link>https://openai.com/index/formal-math</link>
      <description>We built a neural theorem prover for Lean that learned to solve a variety of challenging high-school olympiad problems, including problems from the AMC12 and AIME competitions, as well as two problems adapted from the IMO.</description>
      <guid isPermaLink="true">https://openai.com/index/formal-math</guid>
      <category>Research</category>
      <pubDate>Wed, 02 Feb 2022 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>A research agenda for assessing the economic impacts of code generation models</title>
      <link>https://openai.com/index/economic-impacts-research</link>
      <guid isPermaLink="true">https://openai.com/index/economic-impacts-research</guid>
      <category>Research</category>
      <pubDate>Thu, 03 Mar 2022 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Hierarchical text-conditional image generation with CLIP latents</title>
      <link>https://openai.com/index/hierarchical-text-conditional-image-generation-with-clip-latents</link>
      <guid isPermaLink="true">https://openai.com/index/hierarchical-text-conditional-image-generation-with-clip-latents</guid>
      <category>Research</category>
      <pubDate>Wed, 13 Apr 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Teaching models to express their uncertainty in words</title>
      <link>https://openai.com/index/teaching-models-to-express-their-uncertainty-in-words</link>
      <guid isPermaLink="true">https://openai.com/index/teaching-models-to-express-their-uncertainty-in-words</guid>
      <category>Research</category>
      <pubDate>Sat, 28 May 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Techniques for training large neural networks</title>
      <link>https://openai.com/index/techniques-for-training-large-neural-networks</link>
      <description>Large neural networks are at the core of many recent advances in AI, but training them is a difficult engineering and research challenge which requires orchestrating a cluster of GPUs to perform a single synchronized calculation.</description>
      <guid isPermaLink="true">https://openai.com/index/techniques-for-training-large-neural-networks</guid>
      <category>Research</category>
      <pubDate>Thu, 09 Jun 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evolution through large models</title>
      <link>https://openai.com/index/evolution-through-large-models</link>
      <guid isPermaLink="true">https://openai.com/index/evolution-through-large-models</guid>
      <category>Research</category>
      <pubDate>Fri, 17 Jun 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning to play Minecraft with Video PreTraining</title>
      <link>https://openai.com/index/vpt</link>
      <description>We trained a neural network to play Minecraft by Video PreTraining (VPT) on a massive unlabeled video dataset of human Minecraft play, while using only a small amount of labeled contractor data. With fine-tuning, our model can learn to craft diamond tools, a task that usually takes proficient humans over 20 minutes (24,000 actions). Our model uses the native human interface of keypresses and mouse movements, making it quite general, and represents a step towards general computer-using agents.</description>
      <guid isPermaLink="true">https://openai.com/index/vpt</guid>
      <category>Research</category>
      <pubDate>Thu, 23 Jun 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>DALL·E 2 pre-training mitigations</title>
      <link>https://openai.com/index/dall-e-2-pre-training-mitigations</link>
      <description>In order to share the magic of DALL·E 2 with a broad audience, we needed to reduce the risks associated with powerful image generation models. To this end, we put various guardrails in place to prevent generated images from violating our content policy.</description>
      <guid isPermaLink="true">https://openai.com/index/dall-e-2-pre-training-mitigations</guid>
      <category>Research</category>
      <pubDate>Tue, 28 Jun 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Efficient training of language models to fill in the middle</title>
      <link>https://openai.com/index/efficient-training-of-language-models-to-fill-in-the-middle</link>
      <guid isPermaLink="true">https://openai.com/index/efficient-training-of-language-models-to-fill-in-the-middle</guid>
      <category>Research</category>
      <pubDate>Thu, 28 Jul 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing Whisper</title>
      <link>https://openai.com/index/whisper</link>
      <description>We’ve trained and are open-sourcing a neural net called Whisper that approaches human level robustness and accuracy on English speech recognition.</description>
      <guid isPermaLink="true">https://openai.com/index/whisper</guid>
      <category>Research</category>
      <pubDate>Wed, 21 Sep 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Scaling laws for reward model overoptimization</title>
      <link>https://openai.com/index/scaling-laws-for-reward-model-overoptimization</link>
      <guid isPermaLink="true">https://openai.com/index/scaling-laws-for-reward-model-overoptimization</guid>
      <category>Research</category>
      <pubDate>Wed, 19 Oct 2022 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Point-E: A system for generating 3D point clouds from complex prompts</title>
      <link>https://openai.com/index/point-e</link>
      <guid isPermaLink="true">https://openai.com/index/point-e</guid>
      <category>Research</category>
      <pubDate>Fri, 16 Dec 2022 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-4</title>
      <link>https://openai.com/index/gpt-4-research</link>
      <description>We’ve created GPT-4, the latest milestone in OpenAI’s effort in scaling up deep learning. GPT-4 is a large multimodal model (accepting image and text inputs, emitting text outputs) that, while less capable than humans in many real-world scenarios, exhibits human-level performance on various professional and academic benchmarks.</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-4-research</guid>
      <category>Research</category>
      <pubDate>Tue, 14 Mar 2023 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPTs are GPTs: An early look at the labor market impact potential of large language models</title>
      <link>https://openai.com/index/gpts-are-gpts</link>
      <guid isPermaLink="true">https://openai.com/index/gpts-are-gpts</guid>
      <category>Research</category>
      <pubDate>Fri, 17 Mar 2023 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Democratic inputs to AI</title>
      <link>https://openai.com/index/democratic-inputs-to-ai</link>
      <description>Our nonprofit organization, OpenAI, Inc., is launching a program to award ten $100,000 grants to fund experiments in setting up a democratic process for deciding what rules AI systems should follow, within the bounds defined by the law.</description>
      <guid isPermaLink="true">https://openai.com/index/democratic-inputs-to-ai</guid>
      <category>Research</category>
      <pubDate>Thu, 25 May 2023 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improving mathematical reasoning with process supervision</title>
      <link>https://openai.com/index/improving-mathematical-reasoning-with-process-supervision</link>
      <description>We've trained a model to achieve a new state-of-the-art in mathematical problem solving by rewarding each correct step of reasoning (“process supervision”) instead of simply rewarding the correct final answer (“outcome supervision”). In addition to boosting performance relative to outcome supervision, process supervision also has an important alignment benefit: it directly trains the model to produce a chain-of-thought that is endorsed by humans.</description>
      <guid isPermaLink="true">https://openai.com/index/improving-mathematical-reasoning-with-process-supervision</guid>
      <category>Research</category>
      <pubDate>Wed, 31 May 2023 07:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Building an early warning system for LLM-aided biological threat creation</title>
      <link>https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation</link>
      <description>We’re developing a blueprint for evaluating the risk that a large language model (LLM) could aid someone in creating a biological threat. In an evaluation involving both biology experts and students, we found that GPT-4 provides at most a mild uplift in biological threat creation accuracy. While this uplift is not large enough to be conclusive, our finding is a starting point for continued research and community deliberation. </description>
      <guid isPermaLink="true">https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation</guid>
      <category>Research</category>
      <pubDate>Wed, 31 Jan 2024 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Video generation models as world simulators</title>
      <link>https://openai.com/index/video-generation-models-as-world-simulators</link>
      <description>We explore large-scale training of generative models on video data. Specifically, we train text-conditional diffusion models jointly on videos and images of variable durations, resolutions and aspect ratios. We leverage a transformer architecture that operates on spacetime patches of video and image latent codes. Our largest model, Sora, is capable of generating a minute of high fidelity video. Our results suggest that scaling video generation models is a promising path towards building general purpose simulators of the physical world.</description>
      <guid isPermaLink="true">https://openai.com/index/video-generation-models-as-world-simulators</guid>
      <category>Research</category>
      <pubDate>Thu, 15 Feb 2024 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions</title>
      <link>https://openai.com/index/the-instruction-hierarchy</link>
      <description>Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.</description>
      <guid isPermaLink="true">https://openai.com/index/the-instruction-hierarchy</guid>
      <category>Research</category>
      <pubDate>Fri, 19 Apr 2024 19:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Understanding the source of what we see and hear online</title>
      <link>https://openai.com/index/understanding-the-source-of-what-we-see-and-hear-online</link>
      <description>Today we’re introducing new technology to help researchers identify content created by our tools and joining the Coalition for Content Provenance and Authenticity Steering Committee to promote industry standards. </description>
      <guid isPermaLink="true">https://openai.com/index/understanding-the-source-of-what-we-see-and-hear-online</guid>
      <category>Research</category>
      <pubDate>Tue, 07 May 2024 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Hello GPT-4o</title>
      <link>https://openai.com/index/hello-gpt-4o</link>
      <description>We’re announcing GPT-4 Omni, our new flagship model which can reason across audio, vision, and text in real time.</description>
      <guid isPermaLink="true">https://openai.com/index/hello-gpt-4o</guid>
      <category>Research</category>
      <pubDate>Mon, 13 May 2024 10:05:00 +0000</pubDate>
    </item>
    <item>
      <title>Extracting Concepts from GPT-4</title>
      <link>https://openai.com/index/extracting-concepts-from-gpt-4</link>
      <description>Using new techniques for scaling sparse autoencoders, we automatically identified 16 million patterns in GPT-4's computations.</description>
      <guid isPermaLink="true">https://openai.com/index/extracting-concepts-from-gpt-4</guid>
      <category>Research</category>
      <pubDate>Thu, 06 Jun 2024 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Consistency Models</title>
      <link>https://openai.com/index/consistency-models</link>
      <description>Diffusion models have significantly advanced the fields of image, audio, and video generation, but they depend on an iterative sampling process that causes slow generation.</description>
      <guid isPermaLink="true">https://openai.com/index/consistency-models</guid>
      <category>Research</category>
      <pubDate>Thu, 20 Jun 2024 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improved Techniques for Training Consistency Models</title>
      <link>https://openai.com/index/improved-techniques-for-training-consistency-models</link>
      <description>Consistency models are a nascent family of generative models that can sample high quality data in one step without the need for adversarial training. </description>
      <guid isPermaLink="true">https://openai.com/index/improved-techniques-for-training-consistency-models</guid>
      <category>Research</category>
      <pubDate>Thu, 20 Jun 2024 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>A Holistic Approach to Undesired Content Detection in the Real World</title>
      <link>https://openai.com/index/a-holistic-approach-to-undesired-content-detection-in-the-real-world</link>
      <description>We present a holistic approach to building a robust and useful natural language classification system for real-world content moderation.</description>
      <guid isPermaLink="true">https://openai.com/index/a-holistic-approach-to-undesired-content-detection-in-the-real-world</guid>
      <category>Research</category>
      <pubDate>Thu, 20 Jun 2024 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI and Los Alamos National Laboratory announce research partnership</title>
      <link>https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together</link>
      <description>OpenAI and Los Alamos National Laboratory are working to develop safety evaluations to assess and measure biological capabilities and risks associated with frontier models.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together</guid>
      <category>Research</category>
      <pubDate>Wed, 10 Jul 2024 06:30:00 +0000</pubDate>
    </item>
    <item>
      <title>Prover-Verifier Games improve legibility of language model outputs</title>
      <link>https://openai.com/index/prover-verifier-games-improve-legibility</link>
      <description>Discover how prover-verifier games improve the legibility of language model outputs, making AI solutions clearer, easier to verify, and more trustworthy for both humans and machines.</description>
      <guid isPermaLink="true">https://openai.com/index/prover-verifier-games-improve-legibility</guid>
      <category>Research</category>
      <pubDate>Wed, 17 Jul 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-4o mini: advancing cost-efficient intelligence</title>
      <link>https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence</link>
      <description>Introducing the most cost-efficient small model in the market</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence</guid>
      <category>Research</category>
      <pubDate>Thu, 18 Jul 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improving Model Safety Behavior with Rule-Based Rewards</title>
      <link>https://openai.com/index/improving-model-safety-behavior-with-rule-based-rewards</link>
      <description>We've developed and applied a new method leveraging Rule-Based Rewards (RBRs) that aligns models to behave safely without extensive human data collection.</description>
      <guid isPermaLink="true">https://openai.com/index/improving-model-safety-behavior-with-rule-based-rewards</guid>
      <category>Research</category>
      <pubDate>Wed, 24 Jul 2024 09:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-4o System Card External Testers Acknowledgements</title>
      <link>https://openai.com/index/gpt-4o-system-card/external-testers-acknowledgements</link>
      <description>GPT-4o system card external testers acknowledgements</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-4o-system-card/external-testers-acknowledgements</guid>
      <category>Research</category>
      <pubDate>Thu, 08 Aug 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing SWE-bench Verified</title>
      <link>https://openai.com/index/introducing-swe-bench-verified</link>
      <description>We’re releasing a human-validated subset of SWE-bench that more reliably evaluates AI models’ ability to solve real-world software issues.</description>
      <guid isPermaLink="true">https://openai.com/index/introducing-swe-bench-verified</guid>
      <category>Research</category>
      <pubDate>Tue, 13 Aug 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI o1 System Card External Testers Acknowledgements</title>
      <link>https://openai.com/index/openai-o1-system-card/external-testers-acknowledgements</link>
      <description>OpenAI o1 system card external testers acknowledgements</description>
      <guid isPermaLink="true">https://openai.com/index/openai-o1-system-card/external-testers-acknowledgements</guid>
      <category>Research</category>
      <pubDate>Thu, 12 Sep 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI o1-mini</title>
      <link>https://openai.com/index/openai-o1-mini-advancing-cost-efficient-reasoning</link>
      <description>Advancing cost-efficient reasoning</description>
      <guid isPermaLink="true">https://openai.com/index/openai-o1-mini-advancing-cost-efficient-reasoning</guid>
      <category>Research</category>
      <pubDate>Thu, 12 Sep 2024 10:01:00 +0000</pubDate>
    </item>
    <item>
      <title>Learning to reason with LLMs</title>
      <link>https://openai.com/index/learning-to-reason-with-llms</link>
      <description>We are introducing OpenAI o1, a new large language model trained with reinforcement learning to perform complex reasoning. o1 thinks before it answers—it can produce a long internal chain of thought before responding to the user.</description>
      <guid isPermaLink="true">https://openai.com/index/learning-to-reason-with-llms</guid>
      <category>Research</category>
      <pubDate>Thu, 12 Sep 2024 10:02:00 +0000</pubDate>
    </item>
    <item>
      <title>MLE-bench: Evaluating Machine Learning Agents on Machine Learning Engineering</title>
      <link>https://openai.com/index/mle-bench</link>
      <description>We introduce MLE-bench, a benchmark for measuring how well AI agents perform at machine learning engineering.</description>
      <guid isPermaLink="true">https://openai.com/index/mle-bench</guid>
      <category>Research</category>
      <pubDate>Thu, 10 Oct 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evaluating fairness in ChatGPT</title>
      <link>https://openai.com/index/evaluating-fairness-in-chatgpt</link>
      <description>We've analyzed how ChatGPT responds to users based on their name, using AI research assistants to protect privacy.</description>
      <guid isPermaLink="true">https://openai.com/index/evaluating-fairness-in-chatgpt</guid>
      <category>Research</category>
      <pubDate>Tue, 15 Oct 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Simplifying, stabilizing, and scaling continuous-time consistency models</title>
      <link>https://openai.com/index/simplifying-stabilizing-and-scaling-continuous-time-consistency-models</link>
      <description>We’ve simplified, stabilized, and scaled continuous-time consistency models, achieving comparable sample quality to leading diffusion models, while using only two sampling steps.</description>
      <guid isPermaLink="true">https://openai.com/index/simplifying-stabilizing-and-scaling-continuous-time-consistency-models</guid>
      <category>Research</category>
      <pubDate>Wed, 23 Oct 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing SimpleQA</title>
      <link>https://openai.com/index/introducing-simpleqa</link>
      <description>A factuality benchmark called SimpleQA that measures the ability for language models to answer short, fact-seeking questions.</description>
      <guid isPermaLink="true">https://openai.com/index/introducing-simpleqa</guid>
      <category>Research</category>
      <pubDate>Wed, 30 Oct 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Advancing red teaming with people and AI </title>
      <link>https://openai.com/index/advancing-red-teaming-with-people-and-ai</link>
      <description>Advancing red teaming with people and AI </description>
      <guid isPermaLink="true">https://openai.com/index/advancing-red-teaming-with-people-and-ai</guid>
      <category>Research</category>
      <pubDate>Thu, 21 Nov 2024 10:30:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI o1 System Card</title>
      <link>https://openai.com/index/openai-o1-system-card</link>
      <description>This report outlines the safety work carried out prior to releasing OpenAI o1 and o1-mini, including external red teaming and frontier risk evaluations according to our Preparedness Framework.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-o1-system-card</guid>
      <category>Research</category>
      <pubDate>Thu, 05 Dec 2024 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Trading inference-time compute for adversarial robustness</title>
      <link>https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness</link>
      <description>Trading Inference-Time Compute for Adversarial Robustness</description>
      <guid isPermaLink="true">https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness</guid>
      <category>Research</category>
      <pubDate>Wed, 22 Jan 2025 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Computer-Using Agent</title>
      <link>https://openai.com/index/computer-using-agent</link>
      <description>A universal interface for AI to interact with the digital world.</description>
      <guid isPermaLink="true">https://openai.com/index/computer-using-agent</guid>
      <category>Research</category>
      <pubDate>Thu, 23 Jan 2025 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI o3-mini System Card</title>
      <link>https://openai.com/index/o3-mini-system-card</link>
      <description>This report outlines the safety work carried out for the OpenAI o3-mini model, including safety evaluations, external red teaming, and Preparedness Framework evaluations.</description>
      <guid isPermaLink="true">https://openai.com/index/o3-mini-system-card</guid>
      <category>Research</category>
      <pubDate>Fri, 31 Jan 2025 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>OpenAI o3-mini</title>
      <link>https://openai.com/index/openai-o3-mini</link>
      <description>Pushing the frontier of cost-effective reasoning.</description>
      <guid isPermaLink="true">https://openai.com/index/openai-o3-mini</guid>
      <category>Research</category>
      <pubDate>Fri, 31 Jan 2025 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing deep research</title>
      <link>https://openai.com/index/introducing-deep-research</link>
      <description>An agent that uses reasoning to synthesize large amounts of online information and complete multi-step research tasks for you. Available to Pro users today, Plus and Team next. 

</description>
      <guid isPermaLink="true">https://openai.com/index/introducing-deep-research</guid>
      <category>Research</category>
      <pubDate>Sun, 02 Feb 2025 16:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Why language models hallucinate</title>
      <link>https://openai.com/index/why-language-models-hallucinate</link>
      <description>OpenAI’s new research explains why language models hallucinate. The findings show how improved evaluations can enhance AI reliability, honesty, and safety.</description>
      <guid isPermaLink="true">https://openai.com/index/why-language-models-hallucinate</guid>
      <category>Research</category>
      <pubDate>Fri, 05 Sep 2025 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>How people are using ChatGPT</title>
      <link>https://openai.com/index/how-people-are-using-chatgpt</link>
      <description>New research from the largest study of ChatGPT use shows how the tool creates economic value through both personal and professional use. Adoption is broadening beyond early users, closing gaps and making AI a part of everyday life.</description>
      <guid isPermaLink="true">https://openai.com/index/how-people-are-using-chatgpt</guid>
      <category>Research</category>
      <pubDate>Mon, 15 Sep 2025 03:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Sora 2 is here </title>
      <link>https://openai.com/index/sora-2</link>
      <description>Our latest video generation model is more physically accurate, realistic, and controllable than prior systems. It also features synchronized dialogue and sound effects. Create with it in the new Sora app.</description>
      <guid isPermaLink="true">https://openai.com/index/sora-2</guid>
      <category>Research</category>
      <pubDate>Tue, 30 Sep 2025 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Defining and evaluating political bias in LLMs</title>
      <link>https://openai.com/index/defining-and-evaluating-political-bias-in-llms</link>
      <description>Learn how OpenAI evaluates political bias in ChatGPT through new real-world testing methods that improve objectivity and reduce bias.</description>
      <guid isPermaLink="true">https://openai.com/index/defining-and-evaluating-political-bias-in-llms</guid>
      <category>Research</category>
      <pubDate>Thu, 09 Oct 2025 13:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing IndQA</title>
      <link>https://openai.com/index/introducing-indqa</link>
      <description>OpenAI introduces IndQA, a new benchmark for evaluating AI systems in Indian languages. Built with domain experts, IndQA tests cultural understanding and reasoning across 12 languages and 10 knowledge areas.</description>
      <guid isPermaLink="true">https://openai.com/index/introducing-indqa</guid>
      <category>Research</category>
      <pubDate>Mon, 03 Nov 2025 22:30:00 +0000</pubDate>
    </item>
    <item>
      <title>Understanding neural networks through sparse circuits</title>
      <link>https://openai.com/index/understanding-neural-networks-through-sparse-circuits</link>
      <description>OpenAI is exploring mechanistic interpretability to understand how neural networks reason. Our new sparse model approach could make AI systems more transparent and support safer, more reliable behavior.</description>
      <guid isPermaLink="true">https://openai.com/index/understanding-neural-networks-through-sparse-circuits</guid>
      <category>Research</category>
      <pubDate>Thu, 13 Nov 2025 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>How evals drive the next chapter in AI for businesses</title>
      <link>https://openai.com/index/evals-drive-next-chapter-of-ai</link>
      <description>Learn how evals help businesses define, measure, and improve AI performance—reducing risk, boosting productivity, and driving strategic advantage.</description>
      <guid isPermaLink="true">https://openai.com/index/evals-drive-next-chapter-of-ai</guid>
      <category>Research</category>
      <pubDate>Wed, 19 Nov 2025 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Early experiments in accelerating science with GPT-5</title>
      <link>https://openai.com/index/accelerating-science-gpt-5</link>
      <description>OpenAI introduces the first research cases showing how GPT-5 accelerates scientific progress across math, physics, biology, and computer science. Explore how AI and researchers collaborate to generate proofs, uncover new insights, and reshape the pace of discovery.</description>
      <guid isPermaLink="true">https://openai.com/index/accelerating-science-gpt-5</guid>
      <category>Research</category>
      <pubDate>Thu, 20 Nov 2025 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>How confessions can keep language models honest</title>
      <link>https://openai.com/index/how-confessions-can-keep-language-models-honest</link>
      <description>OpenAI researchers are testing “confessions,” a method that trains models to admit when they make mistakes or act undesirably, helping improve AI honesty, transparency, and trust in model outputs.</description>
      <guid isPermaLink="true">https://openai.com/index/how-confessions-can-keep-language-models-honest</guid>
      <category>Research</category>
      <pubDate>Wed, 03 Dec 2025 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Measuring AI’s capability to accelerate biological research</title>
      <link>https://openai.com/index/accelerating-biological-research-in-the-wet-lab</link>
      <description>OpenAI introduces a real-world evaluation framework to measure how AI can accelerate biological research in the wet lab. Using GPT-5 to optimize a molecular cloning protocol, the work explores both the promise and risks of AI-assisted experimentation.</description>
      <guid isPermaLink="true">https://openai.com/index/accelerating-biological-research-in-the-wet-lab</guid>
      <category>Research</category>
      <pubDate>Tue, 16 Dec 2025 08:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evaluating AI’s ability to perform scientific research tasks</title>
      <link>https://openai.com/index/frontierscience</link>
      <description>OpenAI introduces FrontierScience, a benchmark testing AI reasoning in physics, chemistry, and biology to measure progress toward real scientific research.</description>
      <guid isPermaLink="true">https://openai.com/index/frontierscience</guid>
      <category>Research</category>
      <pubDate>Tue, 16 Dec 2025 09:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Evaluating chain-of-thought monitorability</title>
      <link>https://openai.com/index/evaluating-chain-of-thought-monitorability</link>
      <description>OpenAI introduces a new framework and evaluation suite for chain-of-thought monitorability, covering 13 evaluations across 24 environments. Our findings show that monitoring a model’s internal reasoning is far more effective than monitoring outputs alone, offering a promising path toward scalable control as AI systems grow more capable.</description>
      <guid isPermaLink="true">https://openai.com/index/evaluating-chain-of-thought-monitorability</guid>
      <category>Research</category>
      <pubDate>Thu, 18 Dec 2025 12:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-5 lowers the cost of cell-free protein synthesis</title>
      <link>https://openai.com/index/gpt-5-lowers-protein-synthesis-cost</link>
      <description>An autonomous lab combining OpenAI’s GPT-5 with Ginkgo Bioworks’ cloud automation cut cell-free protein synthesis costs by 40% through closed-loop experimentation.</description>
      <guid isPermaLink="true">https://openai.com/index/gpt-5-lowers-protein-synthesis-cost</guid>
      <category>Research</category>
      <pubDate>Thu, 05 Feb 2026 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>GPT-5.2 derives a new result in theoretical physics</title>
      <link>https://openai.com/index/new-result-theoretical-physics</link>
      <description>A new preprint shows GPT-5.2 proposing a new formula for a gluon amplitude, later formally proved and verified by OpenAI and academic collaborators.</description>
      <guid isPermaLink="true">https://openai.com/index/new-result-theoretical-physics</guid>
      <category>Research</category>
      <pubDate>Fri, 13 Feb 2026 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Introducing EVMbench</title>
      <link>https://openai.com/index/introducing-evmbench</link>
      <description>OpenAI and Paradigm introduce EVMbench, a benchmark evaluating AI agents’ ability to detect, patch, and exploit high-severity smart contract vulnerabilities.</description>
      <guid isPermaLink="true">https://openai.com/index/introducing-evmbench</guid>
      <category>Research</category>
      <pubDate>Wed, 18 Feb 2026 00:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Our First Proof submissions</title>
      <link>https://openai.com/index/first-proof-submissions</link>
      <description>We share our AI model’s proof attempts for the First Proof math challenge, testing research-grade reasoning on expert-level problems.</description>
      <guid isPermaLink="true">https://openai.com/index/first-proof-submissions</guid>
      <category>Research</category>
      <pubDate>Fri, 20 Feb 2026 14:30:00 +0000</pubDate>
    </item>
    <item>
      <title>Why we no longer evaluate SWE-bench Verified</title>
      <link>https://openai.com/index/why-we-no-longer-evaluate-swe-bench-verified</link>
      <description>SWE-bench Verified is increasingly contaminated and mismeasures frontier coding progress. Our analysis shows flawed tests and training leakage. We recommend SWE-bench Pro.</description>
      <guid isPermaLink="true">https://openai.com/index/why-we-no-longer-evaluate-swe-bench-verified</guid>
      <category>Research</category>
      <pubDate>Mon, 23 Feb 2026 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Extending single-minus amplitudes to gravitons</title>
      <link>https://openai.com/index/extending-single-minus-amplitudes-to-gravitons</link>
      <description>A new preprint extends single-minus amplitudes to gravitons, with GPT-5.2 Pro helping derive and verify nonzero graviton tree amplitudes in quantum gravity.</description>
      <guid isPermaLink="true">https://openai.com/index/extending-single-minus-amplitudes-to-gravitons</guid>
      <category>Research</category>
      <pubDate>Wed, 04 Mar 2026 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Reasoning models struggle to control their chains of thought, and that’s good</title>
      <link>https://openai.com/index/reasoning-models-chain-of-thought-controllability</link>
      <description>OpenAI introduces CoT-Control and finds reasoning models struggle to control their chains of thought, reinforcing monitorability as an AI safety safeguard.</description>
      <guid isPermaLink="true">https://openai.com/index/reasoning-models-chain-of-thought-controllability</guid>
      <category>Research</category>
      <pubDate>Thu, 05 Mar 2026 10:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Improving instruction hierarchy in frontier LLMs</title>
      <link>https://openai.com/index/instruction-hierarchy-challenge</link>
      <description>IH-Challenge trains models to prioritize trusted instructions, improving instruction hierarchy, safety steerability, and resistance to prompt injection attacks.</description>
      <guid isPermaLink="true">https://openai.com/index/instruction-hierarchy-challenge</guid>
      <category>Research</category>
      <pubDate>Tue, 10 Mar 2026 11:00:00 +0000</pubDate>
    </item>
    <item>
      <title>Inside our approach to the Model Spec</title>
      <link>https://openai.com/index/our-approach-to-the-model-spec</link>
      <description>Learn how OpenAI’s Model Spec serves as a public framework for model behavior, balancing safety, user freedom, and accountability as AI systems advance.</description>
      <guid isPermaLink="true">https://openai.com/index/our-approach-to-the-model-spec</guid>
      <category>Research</category>
      <pubDate>Wed, 25 Mar 2026 10:00:00 +0000</pubDate>
    </item>
  </channel>
</rss>
