Llama-3.1 Announcement
We are happy to announce that we have brought up support for Llama-3.1-70B inference on Tenstorrent’s 8-chip systems, the TT-QuietBox and the TT-LoudBox.

The source code for Llama-3.1-70B and other models that are supported is on our GitHub. We have also merged support for Llama-3.1-8B, running on our single-chip n150 card.
Implementation highlights:
- Fractured with 8-way tensor parallelism
- Uses FlashAttention and FlashDecode
- Uses Mixed BF16, BFP8, and BFP4 precision
- Performance was measured in eager mode with tracing disabled
We are working on optimizations which will get us to our target of 20 tokens/second/user. Buy our 8-chip systems (TT-QuietBox and TT-LoudBox) to try Llama-3.1-70B at home on Tenstorrent hardware!
Other articles

Tenstorrent Unveils First Gen Compact AI Accelerator Device
Tenstorrent unveiled its first-generation compact AI accelerator device designed in partnership with Razer™ today at CES 2026

Republic of Cyprus Selects Tenstorrent for AI Innovation
Tenstorrent signed a strategic Memorandum of Understanding (MoU) with the Research and Innovation Foundation of Cyprus

Tenstorrent and PwC Partner to Advance AI Ecosystem Development in Cyprus
Tenstorrent and PwC are entering into a strategic collaboration to accelerate AI ecosystem development in Cyprus