view article Article makeMoE: Implement a Sparse Mixture of Experts Language Model from Scratch May 7, 2024 • 112
view article Article No GPU left behind: Unlocking Efficiency with Co-located vLLM in TRL +4 Jun 3, 2025 • 97
NousResearch/DeepHermes-3-Mistral-24B-Preview Text Generation • 24B • Updated Mar 13, 2025 • 1.91k • 121