PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.

Arcee Trinity Large Technical Report

ArXivSource

Varun Singh, Lucas Krauss, Sami Jaghouar, Matej Sirovatka, Charles Goddard, Fares Obied, Jack Min Ong, Jannik Straube, Fern, Aria Harley, Conner Stewart, Colin Kealty, Maziyar Panahi, Simon Kirsten, Anushka Deshpande, Anneketh Vij, Arthur Bresnu, Pranav Veldurthi, Raghav Ravishankar, Hardik Bishnoi, DatologyAI Team, Arcee AI Team, Prime Intellect Team, Mark McQuade, Johannes Hagemann, Lucas Atkins

cs.LG
cs.CL
|
Feb 19, 2026
5 views

One-line Summary

The Arcee Trinity Large is a 400B parameter sparse model using a novel MoE approach, with successful training on 17 trillion tokens and new load balancing strategies.

Plain-language Overview

Researchers have developed a new large-scale AI model called Arcee Trinity Large, which is part of a family of models designed to efficiently process language. This model uses a technique called Mixture-of-Experts (MoE) to activate only a part of its parameters for each input, making it more efficient. The biggest model in this series, Trinity Large, has 400 billion parameters but activates only 13 billion for each token it processes. The team introduced a new method to balance the workload across different parts of the model, ensuring smooth training. The models were trained on vast amounts of text data and are now available for use.

Technical Details