Topic: lab:deep-mind-ai

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for lab:deep-mind-ai, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time
Current monthLast month2 months ago

← Back to home

  1. Faster sorting algorithms discovered using deep reinforcement learning

    PaperJun 7, 2023NatureDaniel J. Mankowitz, Andrea Michi, Anton Zhernov, Marco Gelpi, Marco Selvi, Alhussein Fawzi

    Fundamental algorithms such as sorting or hashing are used trillions of times on any given day. As demand for computation grows, it has become critical for these algorithms to be as performant as p...

  2. Grandmaster level in StarCraft II using multi-agent reinforcement learning

    PaperOct 30, 2019NatureOriol Vinyals, Igor Babuschkin, Wojciech M. Czarnecki, Michaël Mathieu, Andrew Dudzik, David Silver

    The game of StarCraft II has emerged as a grand challenge for artificial intelligence research owing to its complex, multi-agent, and partially observable environment. Here we introduce AlphaStar, ...

  3. Scaling Language Models: Methods, Analysis & Insights from Training Gopher

    PaperDec 8, 2021arXivJack W. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia Creswell, Nat McAleese, Amy Wu, Eleni Elia, Danilo J. Rezende, Vinyals, Simonyan

    Language modelling provides a step towards intelligent communication systems by harnessing large datasets and expressive models. We provide an analysis of Transformer-based language model architect...

  4. Asynchronous Methods for Deep Reinforcement Learning

    PaperFeb 4, 2016arXivVolodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, Koray Kavukcuoglu

    We propose a conceptually simple and lightweight framework for deep reinforcement learning that uses asynchronous gradient descent for optimization of deep neural network controllers. We present as...

  5. A Generalist Agent

    PaperMay 12, 2022arXivScott Reed, Konrad Zolna, Emilio Parisotto, Sergio Gomez Colmenarejo, Nando de Freitas

    Inspired by progress in large-scale language modeling, we apply a similar approach towards building a single generalist agent beyond the realm of text outputs. The agent, which we refer to as Gato,...

← PreviousPage 3Next →

Top Entities In This Topic

Related Topics

FAQ

What does this lab:deep-mind-ai page rank?

It ranks public content for lab:deep-mind-ai using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to lab:deep-mind-ai topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to lab:deep-mind-ai topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in lab:deep-mind-ai?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to lab:deep-mind-ai topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to lab:deep-mind-ai topic page on Attendemia and is written so it still makes sense without reading other sections on the page.