Topic: deep-learning/from/bytedance-research

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for deep-learning/from/bytedance-research, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. FAN: Fourier Analysis Networks

    PaperOct 26, 2025arxiv.orgYihong Dong, Ge Li, Yongding Tao, Xue Jiang, Kechi Zhang, Jia Li, Jinliang Deng, Jing Su, Jun Zhang, Jingjing Xu

    Despite the remarkable successes of general-purpose neural networks, such as MLPs and Transformers, we find that they exhibit notable shortcomings in modeling and reasoning about periodic phenomena...

  2. Loong: Generating Minute-level Long Videos with Autoregressive Language Models

    PaperApr 2, 2025arxiv.orgYuqing Wang, Tianwei Xiong, Daquan Zhou, Zhijie Lin, Yang Zhao, Bingyi Kang, Jiashi Feng, Xihui Liu

    It is desirable but challenging to generate content-rich long videos in the scale of minutes. Autoregressive large language models (LLMs) have achieved great success in generating coherent and long...

  3. Hyper-Connections

    PaperMar 18, 2025arxiv.orgDefa Zhu, Hongzhi Huang, Zihao Huang, Yutao Zeng, Yunyao Mao, Banggu Wu, Qiyang Min, Xun Zhou

    We present hyper-connections, a simple yet effective method that can serve as an alternative to residual connections. This approach specifically addresses common drawbacks observed in residual conn...

  4. ByteCheckpoint: A Unified Checkpointing System for Large Foundation Model Development

    PaperApr 2, 2025arxiv.orgBorui Wan, Mingji Han, Yiyao Sheng, Yanghua Peng, Haibin Lin, Mofan Zhang, Zhichao Lai, Menghan Yu, Junda Zhang, Zuquan Song, Xin Liu, Chuan Wu

    Checkpointing to preserve training states is crucial during the development of Large Foundation Models (LFMs), for training resumption upon various failures or changes in GPU resources and parallel...

  5. Let the Code LLM Edit Itself When You Edit the Code

    PaperMar 4, 2025arxiv.orgZhenyu He, Jun Zhang, Shengjie Luo, Jingjing Xu, Zhi Zhang, Di He

    In this work, we investigate a typical scenario in code generation where a developer edits existing code in real time and requests a code assistant, e.g., a large language model, to re-predict the ...

  6. ClassDiffusion: More Aligned Personalization Tuning with Explicit Class Guidance

    PaperMar 14, 2025arxiv.orgJiannan Huang, Jun Hao Liew, Hanshu Yan, Yuyang Yin, Yao Zhao, Humphrey Shi, Yunchao Wei

    Recent text-to-image customization works have proven successful in generating images of given concepts by fine-tuning diffusion models on a few examples. However, tuning-based methods inherently te...

← PreviousPage 4Next →

Top Entities In This Topic

Related Topics

FAQ

What does this deep-learning/from/bytedance-research page rank?

It ranks public content for deep-learning/from/bytedance-research using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to deep-learning/from/bytedance-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to deep-learning/from/bytedance-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in deep-learning/from/bytedance-research?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to deep-learning/from/bytedance-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to deep-learning/from/bytedance-research topic page on Attendemia and is written so it still makes sense without reading other sections on the page.