Developers
Learn how to get your models up and running fast on Tenstorrent hardware.
With two open source SDKs, you can get as close to the metal as possible, or let our AI compiler do the work.

Model Support Table
QwQ 32B
QuietBox
LLM
TP=8
DeepSeek R1 Distill Llama 3.3 70B
QuietBox
LLM
TP=8
Llama 3.1 70B
Galaxy
LLM
TP=32
Llama 3.1 70B
QuietBox
LLM
TP=8
Llama 3.2 11B Vision
n300
LLM
TP=2
Qwen 2.5 7B
n300
LLM
TP=2
Qwen 2.5 72B
QuietBox
LLM
TP=8
Falcon 7B
n150
LLM
Falcon 7B
QuietBox
LLM
DP=8
Falcon 7B
Galaxy
LLM
DP=32
Falcon 40B
QuietBox
LLM
TP=8
Llama 3.1 8B
n150
LLM
Llama 3.2 1B
n150
LLM
Llama 3.2 3B
n150
LLM
Mamba 2.8B
n150
LLM
Mistral 7B
n150
LLM
Mixtral 8x7B
QuietBox
LLM
TP=8
ResNet-50
n150
CNN
224 x 224
ResNet-50
n300
CNN
224 x 224
DP=2
ResNet-50
QuietBox
CNN
224 x 224
DP=8
ResNet-50
Galaxy
CNN
224 x 224
DP=32
ViT
n150
CNN
224 x 224
Stable Diffusion 1.4
n150
CNN
512 x 512
YOLOv4
n150
CNN
320 x 320
YOLOv4
n150
CNN
640 x 640
SegFormer Semantic Segmentation
n150
CNN
512 x 512
Stable Diffusion 3.5 medium
n150
CNN
512 x 512
BERT-Large
n150
NLP
Getting started on Tenstorrent
TT-Forge™
TT-Forge™ is Tenstorrent's MLIR-based compiler.
TT-NN™
TT-NN™ is a user-friendly API for running ML workloads on Tenstorrent hardware.
TT-Metalium™
TT-Metalium™ is Tenstorrent’s open source, low level AI hardware SDK.
Looking for other documentation?
Active Bounties
Solve bugs and add features to win cash prizes, and get our open source software to stable releases even faster.
Upcoming Events
Apr 16
Office hours: TT-Forge
Have questions about TT-Forge, Tenstorrent's AI compiler? Ask Vraj Prajapati, an engineer on Tenstorrent's Forge team, at our monthly office hours.
Apr 29
Generative AI Summit
Join us in Santa Clara for the Generative AI Summit. Learn more about how you can get started on Tenstorrent hardware, and try out our developer cloud.
Apr 30
Compilers: Modeling Multi-Device and Scale-Out
As machine learning models grow in complexity, executing them efficiently across multiple devices is crucial for scalability and performance. In this talk, Tenstorrent engineer Tapasvi Patel will explore compiler-based techniques for modeling multi-device execution and scale-out strategies.
Educational Content
Tutorials
Written Tutorials
Bring up LLMs with TTNN
Get guidance on how to bring up high-performance multi-chip models on Tenstorrent hardware using the TT-Metalium stack.
Build a graph of a PyTorch based model
In this tutorial, we walk you through the steps of building a graph of a pytorch model.
Join the Community
Get access to support on anything from setting up new hardware, running models or optimizing your setup, plus the latest on Tenstorrent hardware and software.

Interested in contributing?
Tenstorrent's AI software stack is open source. Getting started is as easy as filing an issue.