Arcee AI has released Trinity-Large-Thinking, the full-fledged version of its flagship language model designed for autonomous agent workflows. The model is available both via the company's API and as an open-source release, with weights downloadable under the Apache 2.0 license, allowing for unrestricted use, modification, and deployment on one's own infrastructure.
A Bit of Backstory
To understand why this release is important, it's worth reviewing the past nine months. That's when Arcee AI made a decision that would define the company's future path: to build serious open models in-house, without relying on third-party foundations. The result was the Trinity series.
First came the compact versions – Trinity Nano and Trinity Mini, with approximately 4.5 billion parameters. Then, at the end of January, Trinity-Large-Preview was released, offering a first public glimpse into the capabilities of the larger model. This was an early, 'instruction-tuned' version: the model followed instructions but lacked a built-in 'thinking' mechanism before providing an answer.
Apparently, the Preview found its audience faster than expected. In the first two months, over 3.37 trillion tokens were processed through a major model aggregator, and the model itself became the most used open model in the US and the fourth most popular worldwide in its respective collection.
What's New in This Version
Trinity-Large-Thinking is more than just a refined version of the Preview. The key difference is that the model now 'thinks' before it answers. Simply put, before generating a result, it goes through an internal reasoning phase – an approach familiar from systems like Trinity-Mini.
Why is this necessary? Primarily, for reliable performance in agentic scenarios. An agent is more than just a chatbot: it's a system that executes multi-step tasks, uses tools (e.g., search, code execution, database queries), maintains long dialogues, and needs to not 'break' on the twentieth turn. This is precisely where the Preview showed weaknesses, and it became the main focus during the development of the new version.
According to the team, Trinity-Large-Thinking significantly outperforms the Preview in several key aspects:
- Robust tool use in multi-turn scenarios;
- Maintaining context throughout long sessions;
- Clear adherence to instructions, even with complex constraints;
- Stable performance in long agentic cycles.
On the PinchBench industry benchmark, which measures model capabilities specifically for agentic tasks, Trinity-Large-Thinking took second place – surpassed only by Opus-4.6, but at a significantly lower cost: around $0.90 per million output tokens, which is about 96% cheaper than its competitor.
Openness: A Position, Not a Marketing Ploy
Arcee AI insists that the model's openness is not just a way to attract developers, but a principled stance. The Apache 2.0 license means that companies and developers can not only use the model but also inspect its architecture, fine-tune it for their own tasks, deploy it locally, and embed it into their own products without restrictions.
In the context of the current debate over who controls the development of AI and how, this truly matters. Many powerful models are proprietary, accessible only via API. Trinity-Large-Thinking is one of the rare cases where a genuinely competitive model is being released into the open with full weights.
What's Next?
The team has outlined several directions for future work. First, the experience gained from creating Trinity-Large will be 'passed down the stack' – meaning it will be used to develop the next versions of Nano and Mini. This is standard practice: first, you train a large model, then you distill its knowledge into a smaller one. The next versions of the smaller models, as part of the Trinity-2 series, are already in the works.
Second, work on Trinity-Large itself is not finished – the model will continue to evolve.
Meanwhile, the Preview isn't going anywhere: it will remain free on one of the aggregators, albeit with less dedicated hardware. More details about the long-term plans for this version are promised to be shared later.
Why This Is Interesting
Open-source frontier-level models are a rarity. Companies that achieve competitive results most often opt for a closed API, as it's easier to monetize and control usage. Arcee AI is making a different bet: giving developers and businesses the opportunity to own the model, rather than just rent access to it.
Time will tell how sustainable this strategy will be. But the mere fact that a competitive model with a 'thinking' mode and support for agentic scenarios is being released as open source at an affordable price is a notable event for the entire AI developer community.