AI Agents

LongCat-Flash-Thinking-2601 Technical Report

MMeituan LongCat TeamAAnchun GuiBBei LiBBingyang TaoBBole ZhouBBorun ChenCChao ZhangCChao ZhangCChen GaoCChen ZhangCChengcheng HanCChenhui YangCChuyu ZhangCCong ChenCCunguang WangDDaoru PanDDefei BuDDengchang ZhaoDDi XiuDDishan LiuDDongyu RuDDunwei TuFFan WuFFengcheng YuanFFengcun LiGGang XuGGuanyu WuGGuoyuan LinHHaibin WangHHansi YangHHao YangHHaonan YanHHaoxiang MaHHaoxing WenHHongyan HaoHHongyin TangHHongyu ZangHHongzhi NiHHui SuJJiacheng ZhangJJiahong ZhouJJiahuan LiJJiaming WangJJian YangJJianfei ZhangJJianhao XuJJianing WangJJiapeng ZhuJJiaqi SunJJiarong ShiJJiarui ZhaoJJingang WangJJinluan YangJJinrui DingJJinwei XiaoJJiyuan HeJJuncan XuKKefeng ZhangKKeheng WangLLi WeiLLianhui MaLLin QiuLLingbing KongLLingchuan LiuLLinsen GuoMMengshen ZhuMMengxia ShenMMingyang ZhuPPeiguang LiPPeng PeiPPengcheng JiaPPengtao ZhangPPeng ZhaoQQi GuQQiong HuangQQiyuan DuanQQuanchi WengRRongxiang WengRRongzhi ZhangRRumei LiSShanglin LeiSShengnan AnSShijun DaiSShuaikang LiuSShuang ZhouSShuo WangSSongyuan ZhaoTTao LiangTTianhao HuTTianze ChenWWei LiuWWei ShiWWei WangWWeifeng TangWWenjie ShiWWenlong ZhuWWentao ChenWWentao ShiXXi SuXXiangcheng LiuXXiandi MaXXiangyu XiXXiangyuan LiuXXiangzhou HuangXXiao LiuXXiaodong CaiXXiaolong ChenXXiaowei ShiXXiaoyu LiXXin ChenXXingchen LiuXXuan HuangXXuezhi CaoXXunliang CaiYYan ChenYYang BaiYYang LiuYYang YangYYang ZhengYYaoming WangYYaoming ZhuYYaqi HuoYYanyu ChenYYaorui ShiYYerui SunYYi ZhangYYihao ChenYYi-Kai ZhangYYifan LuYYifan ZhaoYYitao ZhaiYYongjing YinYYongwei ZhouYYoushao XiaoYYuchuan DaiYYuchen XieYYuchen YuYYufei ZhangYYuhuai WeiYYulei QianYYunfan LiangYYunke ZhaoYYuwei JiangYYuxin BianYYuxin ChenYYuxin LiuYYue XuYYueqing SunZZeyang YuZZhao YangZZhengsheng HuangZZhengyu ChenZZhijian LiuZZhikang XiaZZhimin LinZZhiyuan YaoZZhuofan ChenZZhuowen HanZZijian ZhangZZiran LiZZiwen WangZZiyuan Zhuang
arXiv ID
2601.16725
Published
January 23, 2026
Authors
162
Hugging Face Likes
131
Comments
2

Abstract

We introduce LongCat-Flash-Thinking-2601, a 560-billion-parameter open-source Mixture-of-Experts (MoE) reasoning model with superior agentic reasoning capability. LongCat-Flash-Thinking-2601 achieves state-of-the-art performance among open-source models on a wide range of agentic benchmarks, including agentic search, agentic tool use, and tool-integrated reasoning. Beyond benchmark performance, the model demonstrates strong generalization to complex tool interactions and robust behavior under noisy real-world environments. Its advanced capability stems from a unified training framework that combines domain-parallel expert training with subsequent fusion, together with an end-to-end co-design of data construction, environments, algorithms, and infrastructure spanning from pre-training to post-training. In particular, the model's strong generalization capability in complex tool-use are driven by our in-depth exploration of environment scaling and principled task construction. To optimize long-tailed, skewed generation and multi-turn agentic interactions, and to enable stable training across over 10,000 environments spanning more than 20 domains, we systematically extend our asynchronous reinforcement learning framework, DORA, for stable and efficient large-scale multi-environment training. Furthermore, recognizing that real-world tasks are inherently noisy, we conduct a systematic analysis and decomposition of real-world noise patterns, and design targeted training procedures to explicitly incorporate such imperfections into the training process, resulting in improved robustness for real-world applications. To further enhance performance on complex reasoning tasks, we introduce a Heavy Thinking mode that enables effective test-time scaling by jointly expanding reasoning depth and width through intensive parallel thinking.

Keywords

Mixture-of-Expertsagentic reasoningdomain-parallel expert trainingfusionasynchronous reinforcement learningDORAlong-tailed generationmulti-turn interactionsreal-world noise patternstest-time scalingreasoning depthreasoning widthparallel thinking

More in AI Agents

View all
LongCat-Flash-Thinking-2601 Technical Report | Paperchime