Topic: cs.LG

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for cs.LG, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Scaling Laws for Neural Language Models

    PaperJan 23, 2020arXivJared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, Dario Amodei

    We study empirical scaling laws for language model performance on the cross-entropy loss. The loss scales as a power-law with model size, dataset size, and the amount of compute used for training, ...

  2. Proximal Policy Optimization Algorithms

    PaperJul 20, 2017arXivJohn Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, Oleg Klimov

    We propose a new family of policy gradient methods for reinforcement learning, which alternate between sampling data through interaction with the environment, and optimizing a 'surrogate' objective...

  3. Matching Networks for One Shot Learning

    PaperJun 13, 2016arXivOriol Vinyals, Charles Blundell, Timothy Lillicrap, Koray Kavukcuoglu, Daan Wierstra

    Deep learning algorithms typically require vast amounts of data to achieve high performance, contrasting sharply with human ability to learn new concepts from a single example. We introduce Matchin...

  4. Pointer Networks

    PaperJun 9, 2015arXivOriol Vinyals, Meire Fortunato, Navdeep Jaitly

    We introduce a new neural architecture to learn the conditional probability of an output sequence with elements that are discrete tokens corresponding to positions in an input sequence. Such proble...

  5. Human-level performance in 3D multiplayer games with population-based reinforcement learning

    PaperMay 31, 2019ScienceMax Jaderberg, Wojciech M. Czarnecki, Iain Dunning, Luke Marris, Guy Lever, Antonio Garcia Castaneda, Charles Beattie, Neil C. Rabinowitz, Ari S. Morcos, Avraham Ruderman, Nicolas Sonnerat, Tim Green, Louise Deason, Joel Z. Leibo, David Silver, Demis Hassabis, Koray Kavukcuoglu, Thore Graepel

    Multiplayer video games represent a significant frontier for AI research, requiring real-time, high-dimensional sensory processing, spatial navigation, and team-based coordination. We report an AI ...

  6. Mastering Diverse Domains through World Models

    PaperJan 10, 2023arXivDanijar Hafner, Jurgis Pasukonis, Jimmy Ba, Timothy Lillicrap

    General intelligence requires solving tasks across diverse domains without human intervention. We present DreamerV3, a general and scalable reinforcement learning algorithm that masters a wide rang...

  7. Agent57: Outperforming the Atari Human Benchmark

    PaperMar 31, 2020arXivAdrià Puigdomènech Badia, Bilal Piot, Steven Kapturowski, Pablo Sprechmann, Alex Vitvitskyi, Daniel Guo, Charles Blundell

    Atari 2600 games have been a long-standing benchmark in the reinforcement learning community. While previous algorithms have achieved superhuman performance on average, they consistently fail on a ...

  8. Improving language models by retrieving from trillions of tokens

    PaperDec 8, 2021arXivSebastian Borgeaud, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George van den Driessche, Jean-Baptiste Lespiau, Bogdan Damoc, Aidan Clark, Diego de Las Casas, Aurelia Guy, Jacob Menick, Roman Ring, Tom Hennigan, Saffron Huang, Loren Maggiore, Chris Jones, Albin Cassirer, Andy Brock, Michela Paganini, Geoffrey Irving, Oriol Vinyals, Simon Osindero, Karen Simonyan, Jack W. Rae, Erich Elsen, Laurent Sifre

    We enhance auto-regressive language models by conditioning on document chunks retrieved from a large corpus, based on local similarity with preceding tokens. With a 2 trillion token database, our R...

  9. Large Scale GAN Training for High Fidelity Natural Image Synthesis

    PaperSep 28, 2018arXivAndrew Brock, Jeff Donahue, Karen Simonyan

    Despite recent progress in generative image modeling, successfully generating high-resolution, diverse samples from complex datasets such as ImageNet remains an elusive goal. To this end, we train ...

  10. Learning skillful medium-range global weather forecasting

    PaperNov 14, 2023ScienceRemi Lam, Alvaro Sanchez-Gonzalez, Matthew Willson, Peter Wirnsberger, Meire Fortunato, Alexander Alet, Suman Ravuri, Timo Ewalds, Zachary Eaton-Rosen, Weihua Hu, Alexander Merose, Stephan Hoyer, George Holland, Jacklynn Stott, Oriol Vinyals, Shakir Mohamed, Peter Battaglia

    Global medium-range weather forecasting has long been dominated by massive, compute-intensive numerical weather prediction (NWP) models governed by atmospheric physics equations. We present GraphCa...

  11. Perceiver: General Perception with Iterative Attention

    PaperMar 4, 2021arXivAndrew Jaegle, Felix Gimeno, Andrew Brock, Oriol Vinyals, Andrew Zisserman, Joao Carreira

    Biological systems perceive the world by simultaneously processing high-dimensional inputs from modalities as diverse as vision, audition, and touch. We introduce the Perceiver, an architecture tha...

  12. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures

    PaperFeb 9, 2018arXivLasse Espeholt, Hubert Soyer, Remi Munos, Karen Simonyan, Volodymyr Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, Shane Legg, Koray Kavukcuoglu

    Scaling reinforcement learning algorithms to utilize thousands of machines efficiently is crucial for tackling complex, visually rich environments. We introduce IMPALA (Importance Weighted Actor-Le...

  13. Neural Discrete Representation Learning

    PaperNov 2, 2017arXivAaron van den Oord, Oriol Vinyals, Koray Kavukcuoglu

    Learning useful representations without supervision remains a key challenge in machine learning. We propose the Vector Quantised-Variational AutoEncoder (VQ-VAE), a simple yet powerful generative m...

  14. Human-level control through deep reinforcement learning

    PaperFeb 26, 2015NatureVolodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Demis Hassabis

    We present the first deep learning model to successfully learn control policies directly from high-dimensional sensory input using reinforcement learning. The model is a convolutional neural networ...

  15. Verifiable Intent: Cryptographic Anchoring for Autonomous Agency

    PaperFeb 20, 2026arXivJ. L. Martinez, Sarah Chen, Arjun Nair

    To counter the rise of Agent Hijacking, we propose a framework for 'Verifiable Intent.' This protocol cryptographically anchors the human user's original objective to every sub-task generated by a ...

  16. MEG-XL: Data-Efficient Brain-to-Text via Long-Context Pre-Training

    PaperFeb 12, 2026arXivDulhan Jayalath, Oiwi Parker Jones

    Decoding natural language from non-invasive brain recordings like Magnetoencephalography (MEG) remains a significant challenge due to the low signal-to-noise ratio and the scarcity of paired brain-...

  17. High-accuracy sampling for diffusion models and log-concave distributions

    PaperFeb 1, 2026arXivFan Chen, Sinho Chewi, Constantinos Daskalakis, Alexander Rakhlin

    We present algorithms for diffusion model sampling which obtain δ-error in polylog(1/δ) steps, given access to eO(δ)-accurate score estimates in L2. This is an exponential improvement over all prev...

  18. Faster sorting algorithms discovered using deep reinforcement learning

    PaperJun 7, 2023NatureDaniel J. Mankowitz, Andrea Michi, Anton Zhernov, Marco Gelpi, Marco Selvi, Alhussein Fawzi

    Fundamental algorithms such as sorting or hashing are used trillions of times on any given day. As demand for computation grows, it has become critical for these algorithms to be as performant as p...

← PreviousPage 2Next →

Top Entities In This Topic

Related Topics

FAQ

What does this cs.LG page rank?

It ranks public content for cs.LG using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in cs.LG?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.