Topic: Machine Learning

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for Machine Learning, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Diversity-Guided MLP Reduction for Efficient Large Vision Transformers

    PaperSep 22, 2025arxiv.orgChengchao Shen, Hourun Zhu, Gongfan Fang, Jianxin Wang, Xinchao Wang

    Transformer models achieve excellent scaling property, where the performance is improved with the increment of model capacity. However, large-scale model parameters lead to an unaffordable cost of ...

  2. Intriguing properties of neural networks

    PaperFeb 19, 2014arxiv.orgChristian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, Rob Fergus

    Deep neural networks are highly expressive models that have recently achieved state of the art performance on speech and visual recognition tasks. While their expressiveness is the reason they succ...

  3. Recurrent Neural Network Regularization

    PaperFeb 19, 2015arxiv.orgWojciech Zaremba, Ilya Sutskever, Oriol Vinyals

    We present a simple regularization technique for Recurrent Neural Networks (RNNs) with Long Short-Term Memory (LSTM) units. Dropout, the most successful technique for regularizing neural networks, ...

  4. Addressing the Rare Word Problem in Neural Machine Translation

    PaperMay 30, 2015arxiv.orgMinh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Wojciech Zaremba

    Neural Machine Translation (NMT) is a new approach to machine translation that has shown promising results that are comparable to traditional approaches. A significant weakness in conventional NMT ...

  5. Recurrent Models of Visual Attention

    PaperJun 24, 2014arxiv.orgVolodymyr Mnih, Nicolas Heess, Alex Graves, Koray Kavukcuoglu

    Applying convolutional neural networks to large images is computationally expensive because the amount of computation scales linearly with the number of image pixels. We present a novel recurrent n...

  6. A Neural Conversational Model

    PaperJul 22, 2015arxiv.orgOriol Vinyals, Quoc Le

    Conversational modeling is an important task in natural language understanding and machine intelligence. Although previous approaches exist, they are often restricted to specific domains (e.g., boo...

  7. Visualizing and Understanding Recurrent Networks

    PaperNov 17, 2015arxiv.orgAndrej Karpathy, Justin Johnson, Li Fei-Fei

    Recurrent Neural Networks (RNNs), and specifically a variant with Long Short-Term Memory (LSTM), are enjoying renewed interest as a result of successful applications in a wide range of machine lear...

  8. Understanding Neural Networks Through Deep Visualization

    PaperJun 22, 2015arxiv.orgJason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, Hod Lipson

    Recent years have produced great advances in training large, deep neural networks (DNNs), including notable successes in training convolutional neural networks (convnets) to recognize natural image...

  9. Learning Deconvolution Network for Semantic Segmentation

    PaperMay 17, 2015arxiv.orgHyeonwoo Noh, Seunghoon Hong, Bohyung Han

    We propose a novel semantic segmentation algorithm by learning a deconvolution network. We learn the network on top of the convolutional layers adopted from VGG 16-layer net. The deconvolution netw...

  10. Character-Aware Neural Language Models

    PaperDec 1, 2015arxiv.orgYoon Kim, Yacine Jernite, David Sontag, Alexander M. Rush

    We describe a simple neural language model that relies only on character-level inputs. Predictions are still made at the word-level. Our model employs a convolutional neural network (CNN) and a hig...

  11. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

    PaperMar 5, 2016arxiv.orgAnkit Kumar, Ozan Irsoy, Peter Ondruska, Mohit Iyyer, James Bradbury, Ishaan Gulrajani, Victor Zhong, Romain Paulus, Richard Socher

    Most tasks in natural language processing can be cast into question answering (QA) problems over language input. We introduce the dynamic memory network (DMN), a neural network architecture which p...

  12. Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks

    PaperDec 31, 2015arxiv.orgJason Weston, Antoine Bordes, Sumit Chopra, Alexander M. Rush, Bart van Merriënboer, Armand Joulin, Tomas Mikolov

    One long-term goal of machine learning research is to produce methods that are applicable to reasoning and natural language, in particular building an intelligent dialogue agent. To measure progres...

  13. Deep Networks with Stochastic Depth

    PaperJul 28, 2016arxiv.orgGao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, Kilian Weinberger

    Very deep convolutional networks with hundreds of layers have led to significant reductions in error on competitive benchmarks. Although the unmatched expressiveness of the many layers can be highl...

  14. What makes for effective detection proposals?

    PaperAug 1, 2015arxiv.orgJan Hosang, Rodrigo Benenson, Piotr Dollár, Bernt Schiele

    Current top performing object detectors employ detection proposals to guide the search for objects, thereby avoiding exhaustive sliding window search across images. Despite the popularity and wides...

  15. Reading Text in the Wild with Convolutional Neural Networks

    PaperDec 4, 2014arxiv.orgMax Jaderberg, Karen Simonyan, Andrea Vedaldi, Andrew Zisserman

    In this work we present an end-to-end system for text spotting -- localising and recognising text in natural scene images -- and text based image retrieval. This system is based on a region proposa...

← PreviousPage 13Next →

Top Entities In This Topic

Related Topics

FAQ

What does this Machine Learning page rank?

It ranks public content for Machine Learning using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to Machine Learning topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to Machine Learning topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in Machine Learning?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to Machine Learning topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to Machine Learning topic page on Attendemia and is written so it still makes sense without reading other sections on the page.