Topic: cs.CL

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for cs.CL, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Attention Is All You Need

    PaperJun 12, 2017arXivAshish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin

    The dominant sequence transduction models are based on complex recurrent or convolutional neural networks that include an encoder and a decoder. The best performing models also connect the encoder ...

  2. AnyTool: Self-Reflective API Generation for Open-Ended Agentic AI

    PaperJul 22, 2025arXivWei Chen, Yujin Han, Qingwen Bu

    Current Agentic AI systems are constrained by the predefined set of tools provided by developers. We introduce AnyTool, a framework that grants agents the autonomy to dynamically generate, test, an...

  3. TruthfulQA: Measuring How Models Mimic Human Falsehoods

    PaperSep 8, 2021arXivStephanie Lin, Jacob Hilton, Owain Evans

    We propose TruthfulQA, a benchmark to measure whether a language model is truthful in generating answers to questions. The benchmark comprises 817 questions that span 38 categories, including healt...

  4. WebGPT: Browser-assisted question-answering with human feedback

    PaperDec 16, 2021arXivReiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess, John Schulman

    We introduce a method for fine-tuning language models to interact with a text-based web browser to answer open-ended questions. This model, WebGPT, searches the web, navigates through links, and sy...

  5. Learning to summarize from human feedback

    PaperSep 2, 2020arXivNisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, Paul Christiano

    We show that it is possible to significantly improve the quality of text summaries generated by large language models by training them with reinforcement learning from human feedback. We collect a ...

  6. Improving Language Understanding by Generative Pre-Training

    PaperJun 11, 2018OpenAIAlec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever

    Natural language understanding comprises a wide range of diverse tasks such as textual entailment, question answering, semantic similarity assessment, and document classification. Although large un...

  7. Robust Speech Recognition via Large-Scale Weak Supervision

    PaperDec 6, 2022arXivAlec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, Ilya Sutskever

    We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask su...

  8. Language Models are Unsupervised Multitask Learners

    PaperFeb 14, 2019OpenAIAlec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever

    Natural language processing tasks, such as question answering, machine translation, reading comprehension, and summarization, are typically approached with supervised learning on task-specific data...

  9. Learning Transferable Visual Models From Natural Language Supervision

    PaperFeb 26, 2021arXivAlec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever

    State-of-the-art computer vision systems are trained to predict a fixed set of predetermined object categories, restricting their generality. We demonstrate that the simple pre-training task of pre...

  10. Language Models are Few-Shot Learners

    PaperMay 28, 2020arXivTom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei

    Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic i...

  11. Improving language models by retrieving from trillions of tokens

    PaperDec 8, 2021arXivSebastian Borgeaud, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George van den Driessche, Jean-Baptiste Lespiau, Bogdan Damoc, Aidan Clark, Diego de Las Casas, Aurelia Guy, Jacob Menick, Roman Ring, Tom Hennigan, Saffron Huang, Loren Maggiore, Chris Jones, Albin Cassirer, Andy Brock, Michela Paganini, Geoffrey Irving, Oriol Vinyals, Simon Osindero, Karen Simonyan, Jack W. Rae, Erich Elsen, Laurent Sifre

    We enhance auto-regressive language models by conditioning on document chunks retrieved from a large corpus, based on local similarity with preceding tokens. With a 2 trillion token database, our R...

  12. Training Compute-Optimal Large Language Models

    PaperMar 29, 2022arXivJordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Laurent Sifre

    We investigate the optimal model size and number of tokens for training a transformer language model under a given compute budget. We find that current large language models are significantly under...

  13. Flamingo: a Visual Language Model for Few-Shot Learning

    PaperApr 28, 2022arXivJean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Karen Simonyan

    Building models that can be rapidly adapted to novel tasks using only a handful of annotated examples is an open challenge for multimodal machine learning research. We introduce Flamingo, a family ...

  14. Agentic Test-Time Scaling for WebAgents

    PaperFeb 12, 2026arXivNicholas Lee, Lutfi Eren Erdogan, Chris Joseph John, Surya Krishnapillai, Kurt Keutzer, Amir Gholami

    Current WebAgents struggle with long-horizon tasks and complex navigation. We propose an agentic scaling framework that increases compute at test-time through iterative trajectory pruning and rewar...

  15. Gemma: Open Models Based on Gemini Research and Technology

    PaperFeb 21, 2024arXivGemma Team, Google DeepMind

    We introduce Gemma, a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Gemma models are offered in two sizes: a 7 bi...

  16. Improving alignment of dialogue agents via targeted human judgements

    PaperSep 22, 2022arXivAmelia Glaese, Nat McAleese, Maja Trebacz, John Aslanides, Vlad Firoiu, Geoffrey Irving

    We present Sparrow, an information-seeking dialogue agent trained to be more helpful, correct, and harmless compared to prompted language model baselines. We train our model using reinforcement lea...

  17. A-MapReduce: Executing Wide Search via Agentic MapReduce

    PaperFeb 16, 2026arXivMingju Chen, Guibin Zhang, Heng Chang, Yuchen Guo

    Traditional multi-agent systems often struggle with 'search breadth' in unstructured environments, leading to tunnel vision in reasoning. We propose A-MapReduce, a framework that applies the MapRed...

← PreviousPage 1Next →

Top Entities In This Topic

Related Topics

FAQ

What does this cs.CL page rank?

It ranks public content for cs.CL using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to cs.CL topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to cs.CL topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in cs.CL?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to cs.CL topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to cs.CL topic page on Attendemia and is written so it still makes sense without reading other sections on the page.