Ring is a reasoning Mixture-of-Experts (MoE) large language model (LLM) developed by inclusionAI. It is built from or derived from Ling. Its design emphasizes reasoning, efficiency, and modular expert activation. In its “flash” variant (Ring-flash-2.0), it optimizes inference by activating only a subset of experts. It applies reinforcement learning/reasoning optimization techniques. Its architectures and training approaches are tuned to enable efficient and capable reasoning performance. Reasoning-optimized model with reinforcement learning enhancements. Efficient architecture and memory design for large-scale reasoning. If you are located in mainland China, we also provide the model on ModelScope.cn to speed up the download process.
Features
- Mixture-of-Experts (MoE) architecture (activates a subset of experts per input)
- Reasoning-optimized model with reinforcement learning enhancements
- “Thinking” model variant (flash) with sparse expert activation (e.g. 1/32 expert activation)
- High inference throughput (e.g. > 200 tokens/sec under optimized settings)
- Multi-stage training: SFT + RLVR + RLHF
- Efficient architecture and memory design for large-scale reasoning
License
MIT LicenseFollow Ring
You Might Also Like
Gen AI apps are built with MongoDB Atlas
MongoDB Atlas provides built-in vector search and a flexible document model so developers can build, scale, and run gen AI apps without stitching together multiple databases. From LLM integration to semantic search, Atlas simplifies your AI architecture—and it’s free to get started.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of Ring!