Topic: World Model

Track this topic after sign-in.

Short answer

This page shows the most relevant public items for World Model, ranked by trend activity and review signal. Use weekly for fast changes, monthly for more stable patterns, and all-time for evergreen picks.

WeeklyMonthlyAll time

← Back to home

  1. Ultra-Sparse Memory Network

    PaperFeb 6, 2025arxiv.orgZihao Huang, Qiyang Min, Hongzhi Huang, Defa Zhu, Yutao Zeng, Ran Guo, Xun Zhou

    It is widely acknowledged that the performance of Transformer models is logarithmically related to their number of parameters and computational complexity. While approaches like Mixture of Experts ...

  2. Understanding Chain-of-Thought in LLMs through Information Theory

    PaperJul 10, 2025arxiv.orgJean-Francois Ton, Muhammad Faaiz Taufiq, Yang Liu

    Large Language Models (LLMs) have shown impressive performance in complex reasoning tasks through the use of Chain-of-Thought (CoT) reasoning, allowing models to break down problems into manageable...

  3. LSH-MoE: Communication-efficient MoE Training via Locality-Sensitive Hashing

    PaperNov 13, 2024arxiv.orgXiaonan Nie, Qibin Liu, Fangcheng Fu, Shenhan Zhu, Xupeng Miao, Xiaoyang Li, Yang Zhang, Shouda Liu, Bin Cui

    Larger transformer models always perform better on various tasks but require more costs to scale up the model size. To efficiently enlarge models, the mixture-of-experts (MoE) architecture is widel...

  4. SeedEdit: Align Image Re-Generation to Image Editing

    PaperNov 11, 2024arxiv.orgYichun Shi, Peng Wang, Weilin Huang

    We introduce SeedEdit, a diffusion model that is able to revise a given image with any text prompt. In our perspective, the key to such a task is to obtain an optimal balance between maintaining th...

  5. Multi-Reward as Condition for Instruction-based Image Editing

    PaperMar 20, 2025arxiv.orgXin Gu, Ming Li, Libo Zhang, Fan Chen, Longyin Wen, Tiejian Luo, Sijie Zhu

    High-quality training triplets (instruction, original image, edited image) are essential for instruction-based image editing. Predominant training datasets (e.g., InsPix2Pix) are created using text...

  6. Classification Done Right for Vision-Language Pre-Training

    PaperNov 6, 2024arxiv.orgZilong Huang, Qinghao Ye, Bingyi Kang, Jiashi Feng, Haoqi Fan

    We introduce SuperClass, a super simple classification method for vision-language pre-training on image-text data. Unlike its contrastive counterpart CLIP who contrast with a text encoder, SuperCla...

  7. How Far is Video Generation from World Model: A Physical Law Perspective

    PaperJun 22, 2025arxiv.orgBingyi Kang, Yang Yue, Rui Lu, Zhijie Lin, Yang Zhao, Kaixin Wang, Gao Huang, Jiashi Feng

    OpenAI's Sora highlights the potential of video generation for developing world models that adhere to fundamental physical laws. However, the ability of video generation models to discover such law...

  8. AutoKaggle: A Multi-Agent Framework for Autonomous Data Science Competitions

    PaperNov 5, 2024arxiv.orgZiming Li, Qianbo Zang, David Ma, Jiawei Guo, Tuney Zheng, Minghao Liu, Xinyao Niu, Yue Wang, Jian Yang, Jiaheng Liu, Wanjun Zhong, Wangchunshu Zhou, Wenhao Huang, Ge Zhang

    Data science tasks involving tabular data present complex challenges that require sophisticated problem-solving approaches. We propose AutoKaggle, a powerful and user-centric framework that assists...

  9. Why Does the Effective Context Length of LLMs Fall Short?

    PaperOct 24, 2024arxiv.orgChenxin An, Jun Zhang, Ming Zhong, Lei Li, Shansan Gong, Yao Luo, Jingjing Xu, Lingpeng Kong

    Advancements in distributed training and efficient attention mechanisms have significantly expanded the context window sizes of large language models (LLMs). However, recent work reveals that the e...

  10. Reward-Augmented Data Enhances Direct Preference Alignment of LLMs

    PaperMay 11, 2025arxiv.orgShenao Zhang, Zhihan Liu, Boyi Liu, Yufeng Zhang, Yingxiang Yang, Yongfei Liu, Liyu Chen, Tao Sun, Zhaoran Wang

    Preference alignment in Large Language Models (LLMs) has significantly improved their ability to adhere to human instructions and intentions. However, existing direct alignment algorithms primarily...

  11. KOR-Bench: Benchmarking Language Models on Knowledge-Orthogonal Reasoning Tasks

    PaperMar 1, 2025arxiv.orgKaijing Ma, Xinrun Du, Yunran Wang, Haoran Zhang, Zhoufutu Wen, Xingwei Qu, Jian Yang, Jiaheng Liu, Minghao Liu, Xiang Yue, Wenhao Huang, Ge Zhang

    In this paper, we introduce Knowledge-Orthogonal Reasoning (KOR), a concept aimed at minimizing reliance on domain-specific knowledge, enabling more accurate evaluation of models' reasoning abiliti...

  12. FAN: Fourier Analysis Networks

    PaperOct 26, 2025arxiv.orgYihong Dong, Ge Li, Yongding Tao, Xue Jiang, Kechi Zhang, Jia Li, Jinliang Deng, Jing Su, Jun Zhang, Jingjing Xu

    Despite the remarkable successes of general-purpose neural networks, such as MLPs and Transformers, we find that they exhibit notable shortcomings in modeling and reasoning about periodic phenomena...

  13. Loong: Generating Minute-level Long Videos with Autoregressive Language Models

    PaperApr 2, 2025arxiv.orgYuqing Wang, Tianwei Xiong, Daquan Zhou, Zhijie Lin, Yang Zhao, Bingyi Kang, Jiashi Feng, Xihui Liu

    It is desirable but challenging to generate content-rich long videos in the scale of minutes. Autoregressive large language models (LLMs) have achieved great success in generating coherent and long...

  14. Hyper-Connections

    PaperMar 18, 2025arxiv.orgDefa Zhu, Hongzhi Huang, Zihao Huang, Yutao Zeng, Yunyao Mao, Banggu Wu, Qiyang Min, Xun Zhou

    We present hyper-connections, a simple yet effective method that can serve as an alternative to residual connections. This approach specifically addresses common drawbacks observed in residual conn...

  15. MaskBit: Embedding-free Image Generation via Bit Tokens

    PaperDec 8, 2024arxiv.orgMark Weber, Lijun Yu, Qihang Yu, Xueqing Deng, Xiaohui Shen, Daniel Cremers, Liang-Chieh Chen

    Masked transformer models for class-conditional image generation have become a compelling alternative to diffusion models. Typically comprising two stages - an initial VQGAN model for transitioning...

  16. HybridFlow: A Flexible and Efficient RLHF Framework

    PaperOct 2, 2024arxiv.orgGuangming Sheng, Chi Zhang, Zilingfeng Ye, Xibin Wu, Wang Zhang, Ru Zhang, Yanghua Peng, Haibin Lin, Chuan Wu

    Reinforcement Learning from Human Feedback (RLHF) is widely used in Large Language Model (LLM) alignment. Traditional RL can be modeled as a dataflow, where each node represents computation of a ne...

  17. Seed-Music: A Unified Framework for High Quality and Controlled Music Generation

    PaperSep 19, 2024arxiv.orgYe Bai, Haonan Chen, Jitong Chen, Zhuo Chen, Yi Deng, Xiaohong Dong, Lamtharn Hantrakul, Weituo Hao, Qingqing Huang, Zhongyi Huang, Dongya Jia, Feihu La, Duc Le, Bochen Li, Chumin Li, Hui Li, Xingxing Li, Shouda Liu, Wei-Tsung Lu, Yiqing Lu, Andrew Shaw, Janne Spijkervet, Yakun Sun, Bo Wang, Ju-Chiang Wang, Yuping Wang, Yuxuan Wang, Ling Xu, Yifeng Yang, Chao Yao, Shuo Zhang, Yang Zhang, Yilin Zhang, Hang Zhao, Ziyi Zhao, Dejian Zhong, Shicen Zhou, Pei Zou

    We introduce Seed-Music, a suite of music generation systems capable of producing high-quality music with fine-grained style control. Our unified framework leverages both auto-regressive language m...

← PreviousPage 4Next →

Top Entities In This Topic

Related Topics

FAQ

What does this World Model page rank?

It ranks public content for World Model using recent discussion, review, and engagement signals so you can triage faster. This guidance is specific to World Model topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How should I use weekly vs monthly vs all-time?

Use weekly for fast-moving updates, monthly for stable trend confirmation, and all-time for evergreen references. This guidance is specific to World Model topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

How can I discover organizations active in World Model?

Use the linked entities section to jump to labs, companies, and experts connected to this topic and explore their timelines. This guidance is specific to World Model topic page on Attendemia and is written so it still makes sense without reading other sections on the page.

Can I follow this topic for updates?

Yes. Use the follow button on this page to subscribe and track new high-signal activity. This guidance is specific to World Model topic page on Attendemia and is written so it still makes sense without reading other sections on the page.