Topic: Machine Learning

Short answer

This page shows the most relevant public items for Machine Learning, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Diffusion Alignment as Variational Expectation-Maximization

    PaperFeb 13, 2026arXivZijing Ou, Jacob Si, Junyi Zhu, Yingzhen Li

    Diffusion alignment aims to optimize diffusion models for downstream objectives. While existing methods based on RL achieve success, they often suffer from reward over-optimization and mode collaps...

  2. A hyperparameter benchmark of VAE-based methods for scRNA-seq

    PaperFeb 10, 2026bioRxivEduardo da Veiga Beltrame

    This paper presents a systematic benchmark of architectural hyperparameters for variational autoencoder (VAE) methods in single-cell RNA-seq batch integration. The study compares scVI, MrVI, and LD...

  3. High-accuracy sampling for diffusion models and log-concave distributions

    PaperFeb 1, 2026arXivFan Chen, Sinho Chewi, Constantinos Daskalakis, Alexander Rakhlin

    We present algorithms for diffusion model sampling which obtain δ-error in polylog(1/δ) steps, given access to eO(δ)-accurate score estimates in L2. This is an exponential improvement over all prev...

  4. Intriguing properties of neural networks

    PaperFeb 19, 2014arxiv.orgChristian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, Rob Fergus

    Deep neural networks are highly expressive models that have recently achieved state of the art performance on speech and visual recognition tasks. While their expressiveness is the reason they succ...

  5. Recurrent Neural Network Regularization

    PaperFeb 19, 2015arxiv.orgWojciech Zaremba, Ilya Sutskever, Oriol Vinyals

    We present a simple regularization technique for Recurrent Neural Networks (RNNs) with Long Short-Term Memory (LSTM) units. Dropout, the most successful technique for regularizing neural networks, ...

  6. Addressing the Rare Word Problem in Neural Machine Translation

    PaperMay 30, 2015arxiv.orgMinh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Wojciech Zaremba

    Neural Machine Translation (NMT) is a new approach to machine translation that has shown promising results that are comparable to traditional approaches. A significant weakness in conventional NMT ...

  7. Recurrent Models of Visual Attention

    PaperJun 24, 2014arxiv.orgVolodymyr Mnih, Nicolas Heess, Alex Graves, Koray Kavukcuoglu

    Applying convolutional neural networks to large images is computationally expensive because the amount of computation scales linearly with the number of image pixels. We present a novel recurrent n...

  8. A Neural Conversational Model

    PaperJul 22, 2015arxiv.orgOriol Vinyals, Quoc Le

    Conversational modeling is an important task in natural language understanding and machine intelligence. Although previous approaches exist, they are often restricted to specific domains (e.g., boo...

  9. Visualizing and Understanding Recurrent Networks

    PaperNov 17, 2015arxiv.orgAndrej Karpathy, Justin Johnson, Li Fei-Fei

    Recurrent Neural Networks (RNNs), and specifically a variant with Long Short-Term Memory (LSTM), are enjoying renewed interest as a result of successful applications in a wide range of machine lear...

  10. Understanding Neural Networks Through Deep Visualization

    PaperJun 22, 2015arxiv.orgJason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, Hod Lipson

    Recent years have produced great advances in training large, deep neural networks (DNNs), including notable successes in training convolutional neural networks (convnets) to recognize natural image...

  11. Learning Deconvolution Network for Semantic Segmentation

    PaperMay 17, 2015arxiv.orgHyeonwoo Noh, Seunghoon Hong, Bohyung Han

    We propose a novel semantic segmentation algorithm by learning a deconvolution network. We learn the network on top of the convolutional layers adopted from VGG 16-layer net. The deconvolution netw...

  12. Character-Aware Neural Language Models

    PaperDec 1, 2015arxiv.orgYoon Kim, Yacine Jernite, David Sontag, Alexander M. Rush

    We describe a simple neural language model that relies only on character-level inputs. Predictions are still made at the word-level. Our model employs a convolutional neural network (CNN) and a hig...

Related Topics

Awesome List: deep-learning-foundation (115)Deep Learning (115)cs.LG (8)cs.AI (5)Bioinformatics (2)
Machine Learning — trending content and reviews | Attendemia