Topic: company:openai-research

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for company:openai-research, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time
Current monthLast month2 months ago

← Back to home

  1. Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision

    PaperDec 14, 2023arXivCollin Burns, Pavel Izmailov, Jan Hendrik Kirchner, Bowen Baker, Leo Gao, Leopold Aschenbrenner, Yining Chen, Adrien Ecoffet, Manas Joglekar, Jan Leike, Ilya Sutskever, Jeff Wu

    As AI models become increasingly capable, we will eventually face the challenge of superalignment: how can humans supervise AI systems that are much smarter than them? To study this empirically tod...

  2. Consistency Models

    PaperMar 2, 2023arXivYang Song, Prafulla Dhariwal, Mark Chen, Ilya Sutskever

    Diffusion models have achieved significant success in image, audio, and video generation, but they depend on an iterative generation process that causes slow sampling and precludes real-time applic...

  3. Sora: Video generation models as world simulators

    PaperFeb 15, 2024OpenAI Technical ReportTim Brooks, Bill Peebles, Connor Holmes, Will DePue, Yufei Guo, Li Jing, David Schnurr, Joe Taylor, Troy Luhman, Eric Luhman, Clarence Ng, Ricky Wang, Aditya Ramesh

    We explore the large-scale training of generative models on video data. Specifically, we train text-conditional diffusion models jointly on videos and images of highly variable durations, resolutio...

  4. WebGPT: Browser-assisted question-answering with human feedback

    PaperDec 16, 2021arXivReiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess, John Schulman

    We introduce a method for fine-tuning language models to interact with a text-based web browser to answer open-ended questions. This model, WebGPT, searches the web, navigates through links, and sy...

  5. Learning Dexterous In-Hand Manipulation

    PaperJul 30, 2018arXivMarcin Andrychowicz, Bowen Baker, Maciek Chociej, Rafal Jozefowicz, Bob McGrew, Jakub Pachocki, Arthur Petron, Matthias Plappert, Glenn Powell, Alex Ray, Jonas Schneider, Szymon Sidor, Josh Tobin, Peter Welinder, Lilian Weng, Wojciech Zaremba

    We demonstrate that reinforcement learning algorithms can be used to learn highly dexterous, in-hand manipulation policies that successfully transfer to the real world. We train a policy to control...

  6. Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets

    PaperJan 6, 2022arXivAlethea Power, Yuri Burda, Harri Edwards, Igor Babuschkin, Vedant Misra

    We demonstrate a striking phenomenon in the training dynamics of neural networks on small algorithmic datasets: networks that initially severely overfit the training data can, after continued train...

  7. Improved Denoising Diffusion Probabilistic Models

    PaperFeb 18, 2021arXivAlex Nichol, Prafulla Dhariwal

    Denoising diffusion probabilistic models (DDPMs) have recently demonstrated high-quality image generation, but they suffer from notoriously slow sampling times and sub-optimal log-likelihoods. We p...

  8. Dota 2 with Large Scale Deep Reinforcement Learning

    PaperDec 13, 2019arXivChristopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemysław Dębiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, Ilya Sutskever, et al.

    We present OpenAI Five, a system of five neural networks that learned to play the highly complex, imperfect-information esports game Dota 2 entirely through self-play. Dota 2 involves long time hor...

  9. Learning to summarize from human feedback

    PaperSep 2, 2020arXivNisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, Paul Christiano

    We show that it is possible to significantly improve the quality of text summaries generated by large language models by training them with reinforcement learning from human feedback. We collect a ...

  10. Solving Rubik's Cube with a Robot Hand

    PaperOct 15, 2019arXivIlge Akkaya, Marcin Andrychowicz, Maciek Chociej, Mateusz Litwin, Bob McGrew, Arthur Petron, Alex Paino, Matthias Plappert, Glenn Powell, Raphael Ribas, Jonas Schneider, Nikolas Tezak, Peter Welinder, Lilian Weng, Wojciech Zaremba, Lei Zhang

    We demonstrate that models trained only in simulation can be used to solve a manipulation problem of unprecedented complexity on a real robot. We use reinforcement learning to train a policy to sol...

  11. Evaluating Large Language Models Trained on Code

    PaperJul 7, 2021arXivMark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, Alex Ray, Wojciech Zaremba, Ilya Sutskever, et al.

    We introduce Codex, a GPT language model fine-tuned on publicly available code from GitHub, and study its Python code-writing capabilities. A distinct production version of Codex powers GitHub Copi...

  12. Zero-Shot Text-to-Image Generation

    PaperFeb 24, 2021arXivAditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, Ilya Sutskever

    Text-to-image generation has traditionally focused on finding better modeling assumptions for training on a fixed dataset. We describe a simple approach for this task based on a transformer that au...

  13. Scaling Laws for Neural Language Models

    PaperJan 23, 2020arXivJared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, Dario Amodei

    We study empirical scaling laws for language model performance on the cross-entropy loss. The loss scales as a power-law with model size, dataset size, and the amount of compute used for training, ...

  14. Improving Language Understanding by Generative Pre-Training

    PaperJun 11, 2018OpenAIAlec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever

    Natural language understanding comprises a wide range of diverse tasks such as textual entailment, question answering, semantic similarity assessment, and document classification. Although large un...

  15. Robust Speech Recognition via Large-Scale Weak Supervision

    PaperDec 6, 2022arXivAlec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, Ilya Sutskever

    We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask su...

  16. Hierarchical Text-Conditional Image Generation with CLIP Latents

    PaperApr 13, 2022arXivAditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, Mark Chen

    Contrastive models like CLIP have been shown to learn robust representations of images that capture both semantics and style. To leverage these representations for image generation, we propose a tw...

  17. Language Models are Unsupervised Multitask Learners

    PaperFeb 14, 2019OpenAIAlec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever

    Natural language processing tasks, such as question answering, machine translation, reading comprehension, and summarization, are typically approached with supervised learning on task-specific data...

  18. Proximal Policy Optimization Algorithms

    PaperJul 20, 2017arXivJohn Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, Oleg Klimov

    We propose a new family of policy gradient methods for reinforcement learning, which alternate between sampling data through interaction with the environment, and optimizing a 'surrogate' objective...

  19. Learning Transferable Visual Models From Natural Language Supervision

    PaperFeb 26, 2021arXivAlec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever

    State-of-the-art computer vision systems are trained to predict a fixed set of predetermined object categories, restricting their generality. We demonstrate that the simple pre-training task of pre...

← PreviousPage 2Next →

Top Entities In This Topic

Related Topics

FAQ

What does this company:openai-research page rank?

It ranks public content for company:openai-research using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in company:openai-research?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.