AI Research Wiki

Tag: mixture-of-experts

1 item with this tag.

  • Apr 11, 2026

    Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity

    • mixture-of-experts
    • sparsity
    • scaling
    • efficiency

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community