LLaDA2.0: Scaling Up Diffusion Language Models to 100B Paper • 2512.15745 • Published 23 days ago • 77
Every Token Counts: Generalizing 16M Ultra-Long Context in Large Language Models Paper • 2511.23319 • Published Nov 28, 2025 • 22
Every Token Counts: Generalizing 16M Ultra-Long Context in Large Language Models Paper • 2511.23319 • Published Nov 28, 2025 • 22
Every Token Counts: Generalizing 16M Ultra-Long Context in Large Language Models Paper • 2511.23319 • Published Nov 28, 2025 • 22 • 2
QUBE: Enhancing Automatic Heuristic Design via Quality-Uncertainty Balanced Evolution Paper • 2412.20694 • Published Dec 30, 2024
Grove MoE: Towards Efficient and Superior MoE LLMs with Adjugate Experts Paper • 2508.07785 • Published Aug 11, 2025 • 28
MoBE: Mixture-of-Basis-Experts for Compressing MoE-based LLMs Paper • 2508.05257 • Published Aug 7, 2025 • 13
Grove MoE: Towards Efficient and Superior MoE LLMs with Adjugate Experts Paper • 2508.07785 • Published Aug 11, 2025 • 28
Value Residual Learning For Alleviating Attention Concentration In Transformers Paper • 2410.17897 • Published Oct 23, 2024 • 9 • 2
Dynamics of Instruction Tuning: Each Ability of Large Language Models Has Its Own Growth Pace Paper • 2310.19651 • Published Oct 30, 2023
Value Residual Learning For Alleviating Attention Concentration In Transformers Paper • 2410.17897 • Published Oct 23, 2024 • 9