Ming-Flash-Omni: A Sparse, Unified Architecture for Multimodal Perception and Generation Paper • 2510.24821 • Published 4 days ago • 27
Leveraging Large Language Models for Pre-trained Recommender Systems Paper • 2308.10837 • Published Aug 21, 2023 • 1
Professional Agents -- Evolving Large Language Models into Autonomous Experts with Human-Level Competencies Paper • 2402.03628 • Published Feb 6, 2024
Intelligent Virtual Assistants with LLM-based Process Automation Paper • 2312.06677 • Published Dec 4, 2023 • 1
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs Paper • 2503.05139 • Published Mar 7 • 4
A Causal Explainable Guardrails for Large Language Models Paper • 2405.04160 • Published May 7, 2024 • 1
Every Attention Matters: An Efficient Hybrid Architecture for Long-Context Reasoning Paper • 2510.19338 • Published 10 days ago • 101
Every Attention Matters: An Efficient Hybrid Architecture for Long-Context Reasoning Paper • 2510.19338 • Published 10 days ago • 101
Every Attention Matters: An Efficient Hybrid Architecture for Long-Context Reasoning Paper • 2510.19338 • Published 10 days ago • 101 • 4
view article Article Art of Focus: Page-Aware Sparse Attention and Ling 2.0’s Quest for Efficient Context Length Scaling By RichardBian and 19 others • 12 days ago • 14
view article Article Art of Focus: Page-Aware Sparse Attention and Ling 2.0’s Quest for Efficient Context Length Scaling By RichardBian and 19 others • 12 days ago • 14
view article Article Ring-flash-linear-2.0: A Highly Efficient Hybrid Architecture for Test-Time Scaling By RichardBian and 8 others • 23 days ago • 10
view article Article Ring-flash-linear-2.0: A Highly Efficient Hybrid Architecture for Test-Time Scaling By RichardBian and 8 others • 23 days ago • 10