Topic: company:openai-research

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for company:openai-research, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Language Models are Few-Shot Learners

    PaperMay 28, 2020arXivTom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei

    Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic i...

  2. Generative Pretraining from Pixels

    PaperJun 17, 2020OpenAIMark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, Ilya Sutskever

    Inspired by the success of unsupervised representation learning in natural language processing with models like GPT-2, we examine whether similar models can learn useful representations for images....

  3. Jukebox: A Generative Model for Music

    PaperApr 30, 2020arXivPrafulla Dhariwal, Heewoo Jun, Christine Payne, Jong Wook Kim, Alec Radford, Ilya Sutskever

    We introduce Jukebox, a generative model that produces high-fidelity, highly diverse music with singing in the raw audio domain. We model music as a sequence of discrete tokens by using a multi-sca...

  4. OpenAI o1 System Card

    PaperSep 12, 2024OpenAIOpenAI

    We introduce OpenAI o1, a new series of large language models trained with reinforcement learning to perform complex reasoning. o1 models are designed to spend more time thinking before they respon...

  5. Point-E: A System for Generating 3D Point Clouds from Complex Prompts

    PaperDec 16, 2022arXivAlex Nichol, Heewoo Jun, Prafulla Dhariwal, Pamela Mishkin, Mark Chen

    While text-to-image generation has witnessed rapid progress, text-to-3D synthesis remains challenging due to the lack of massive 3D datasets and the complexity of 3D representations. We introduce P...

  6. Trust Region Policy Optimization

    PaperFeb 19, 2015arXivJohn Schulman, Sergey Levine, Pieter Abbeel, Michael Jordan, Philipp Moritz

    We describe an iterative procedure for optimizing policies, with guaranteed monotonic improvement. By making several approximations to the theoretically-justified procedure, we develop a practical ...

  7. GPT-4 Technical Report

    PaperMar 15, 2023arXivOpenAI

    We report the development of GPT-4, a large-scale, multimodal model which can accept image and text inputs and produce text outputs. While less capable than humans in many real-world scenarios, GPT...

  8. Training language models to follow instructions with human feedback

    PaperMar 4, 2022arXivLong Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, Ryan Lowe

    Making language models bigger does not inherently make them better at following a user's intent. For example, large language models can generate outputs that are untruthful, toxic, or simply not he...

  9. OpenAI Gym

    PaperJun 5, 2016arxiv.orgGreg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, Wojciech Zaremba

    OpenAI Gym is a toolkit for reinforcement learning research. It includes a growing collection of benchmark problems that expose a common interface, and a website where people can share their result...

  10. Evolution Strategies as a Scalable Alternative to Reinforcement Learning

    PaperMar 10, 2017arxiv.orgTim Salimans, Jonathan Ho, Xi Chen, Ilya Sutskever

    We explore the use of Evolution Strategies, a class of black box optimization algorithms, as an alternative to popular RL techniques such as Q-learning and Policy Gradients. Experiments on MuJoCo a...

← PreviousPage 3Next →

Top Entities In This Topic

Related Topics

FAQ

What does this company:openai-research page rank?

It ranks public content for company:openai-research using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in company:openai-research?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to company:openai-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.