By using our website, you agree to the collection and processing of your data collected by 3rd party. See GDPR policy
Compact mode

Mixture of Experts V2

Improved MoE with better expert routing and efficiency

Known for Efficient Large Model Scaling

Core Classification

Industry Relevance

Historical Information

Application Domain

Technical Characteristics

Evaluation

Facts

  • Interesting Fact 🤓

    Fascinating trivia or lesser-known information about the algorithm
    • Uses only fraction of parameters per inference

FAQ about Mixture of Experts V2

Contact: [email protected]