LoongTrain: Efficient Training of Long-Sequence LLMs with Head-Context Parallelism Paper • 2406.18485 • Published Jun 26, 2024 • 2
Expert-as-a-Service: Towards Efficient, Scalable, and Robust Large-scale MoE Serving Paper • 2509.17863 • Published Sep 22
BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via Balanced Policy Optimization with Adaptive Clipping Paper • 2510.18927 • Published Oct 21 • 82
ReSpec: Towards Optimizing Speculative Decoding in Reinforcement Learning Systems Paper • 2510.26475 • Published 29 days ago
InternEvo: Efficient Long-sequence Large Language Model Training via Hybrid Parallelism and Redundant Sharding Paper • 2401.09149 • Published Jan 17, 2024 • 1
AMSP: Super-Scaling LLM Training via Advanced Model States Partitioning Paper • 2311.00257 • Published Nov 1, 2023 • 10