Topic: cs.LG

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for cs.LG, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Minimax M2.5: Scaling RL for Industrial-Grade Agentic AI

    PaperFeb 16, 2026arXivMiniMax Research Team

    Training agents for industrial-scale deployment requires extreme stability and data throughput. We present Minimax M2.5, a model trained using a novel asynchronous RL architecture designed to proce...

  2. Fast KV Compaction via Attention Matching

    PaperFeb 18, 2026arXivAdam Zweiger, Xinghong Fu, Han Guo, MIT Team

    Large Language Models struggle with memory overhead during long-context inference due to the linear growth of the Key-Value (KV) cache. We propose Attention Matching (AM), a framework for high-qual...

  3. KLong: Training LLM Agents for Extremely Long-horizon Tasks

    PaperFeb 19, 2026arXivYue Liu, Zhiyuan Hu, Flood Sung

    Current LLM agents frequently fail in tasks requiring hundreds of steps due to error accumulation and context overflow. We introduce KLong, an agentic framework that utilizes 'Trajectory-Splitting ...

  4. Trust Region Policy Optimization

    PaperFeb 19, 2015arXivJohn Schulman, Sergey Levine, Pieter Abbeel, Michael Jordan, Philipp Moritz

    We describe an iterative procedure for optimizing policies, with guaranteed monotonic improvement. By making several approximations to the theoretically-justified procedure, we develop a practical ...

  5. Scaling Language Models: Methods, Analysis & Insights from Training Gopher

    PaperDec 8, 2021arXivJack W. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia Creswell, Nat McAleese, Amy Wu, Eleni Elia, Danilo J. Rezende, Vinyals, Simonyan

    Language modelling provides a step towards intelligent communication systems by harnessing large datasets and expressive models. We provide an analysis of Transformer-based language model architect...

  6. Asynchronous Methods for Deep Reinforcement Learning

    PaperFeb 4, 2016arXivVolodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, Koray Kavukcuoglu

    We propose a conceptually simple and lightweight framework for deep reinforcement learning that uses asynchronous gradient descent for optimization of deep neural network controllers. We present as...

  7. A Generalist Agent

    PaperMay 12, 2022arXivScott Reed, Konrad Zolna, Emilio Parisotto, Sergio Gomez Colmenarejo, Nando de Freitas

    Inspired by progress in large-scale language modeling, we apply a similar approach towards building a single generalist agent beyond the realm of text outputs. The agent, which we refer to as Gato,...

  8. MEM1: A Constant-Memory RL Framework for Long-Horizon Language Agents

    PaperFeb 12, 2026arXivYurong Chen, Yu He, Michael I. Jordan, Fan Yao

    Modern language agents must operate over long-horizon, multi-turn interactions, but most rely on full-context prompting which leads to unbounded memory growth. We introduce MEM1, an end-to-end rein...

← PreviousPage 3Next →

Top Entities In This Topic

Related Topics

FAQ

What does this cs.LG page rank?

It ranks public content for cs.LG using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in cs.LG?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to cs.LG topic page on Attendemia and is written so it still makes sense without reading other sections on the page.