N150 N300 T3K P100 P150 P300C Galaxy 10 min Validated

Download Model from Hugging Face

Download Qwen3-0.6B — the recommended model for Tenstorrent hardware. It's tiny (0.6B parameters), fast, reasoning-capable, and requires no special license agreement. Works reliably on every supported device including N150 and P300c.

Prerequisites

You'll need a Hugging Face access token to download models. If you don't have one:

  1. Go to huggingface.co
  2. Sign up or log in
  3. Navigate to Settings → Access Tokens
  4. Create a new token with read permissions

Starting Fresh?

If you're jumping directly to this lesson, let's verify your setup:

Quick Prerequisite Checks

# Hardware detected?
tt-smi -s

# Python installed?
python3 --version  # Need 3.10+

# hf CLI installed? (included with huggingface-hub)
which hf || pip install huggingface-hub

All checks passed? Continue below to download the model.

If hardware check failed:


Already Authenticated?

Check if you're already logged in to Hugging Face:

hf auth whoami

If this shows your username: You're already authenticated! Skip to Step 3: Download the Model.

If it shows an error: Continue with Step 1 below to authenticate.


Model Already Downloaded?

Check if Qwen3-0.6B is already present:

ls ~/models/Qwen3-0.6B/config.json

If the file exists: Model is already downloaded! You can skip to the next lesson.

Verify download contents:

ls ~/models/Qwen3-0.6B/config.json
ls ~/models/Qwen3-0.6B/model*.safetensors

If you already have Llama-3.1-8B-Instruct:

ls ~/models/Llama-3.1-8B-Instruct/config.json

All files present? You're good to go!

Missing files? Redownload using Step 3 below.


Understanding Model Formats

Qwen3-0.6B uses HuggingFace format only:

The HuggingFace format is the standard for modern models and is compatible with all Tenstorrent inference tools.


Step 1: Set Your Token

First, check if your token is already set:

echo $HF_TOKEN

If you see your token: It's already set! Skip to Step 2: Authenticate.

If it's empty: Set your token using one of these methods:

When you click the button below, you'll be prompted to enter your token securely:

🔑 Enter Your Hugging Face Token

Method 2: Manually in Terminal

export HF_TOKEN=your_token_from_huggingface

Note: This only lasts for your current terminal session. For permanent setup, add it to ~/.bashrc or ~/.zshrc:

echo 'export HF_TOKEN=your_token_from_huggingface' >> ~/.bashrc
source ~/.bashrc

Step 2: Authenticate

Once your token is set, authenticate with Hugging Face:

hf auth login --token "$HF_TOKEN"

✓ Authenticate with Hugging Face
hf auth login --token "$HF_TOKEN"


Step 3: Download Qwen3-0.6B

Tip — custom storage location: All lessons use ~/models as the model directory. If your models live on a larger drive or shared storage, symlink ~/models there once and every lesson will find them automatically:

ln -s /path/to/your/storage ~/models

Already downloaded a model to the HF cache? Symlink the snapshot:

mkdir -p ~/models
ln -sfn "$(ls ~/.cache/huggingface/hub/models--Qwen--Qwen3-0.6B/snapshots/ | tail -1 | xargs -I{} echo ~/.cache/huggingface/hub/models--Qwen--Qwen3-0.6B/snapshots/{})" ~/models/Qwen3-0.6B

Download Qwen3-0.6B — no license gate, no terms to accept, works on all Tenstorrent hardware:

mkdir -p ~/models && hf download Qwen/Qwen3-0.6B \
  --local-dir ~/models/Qwen3-0.6B

⬇️ Download Qwen3-0.6B Model

What Gets Downloaded

The Qwen3-0.6B model includes:

The download typically completes in under a minute on a fast connection.


Optional: Llama-3.1-8B-Instruct (Gated Model)

Note: Meta requires accepting their data license at huggingface.co/meta-llama/Llama-3.1-8B-Instruct before this download will succeed. If you prefer open models or haven't accepted Meta's terms, Qwen3-0.6B is an excellent alternative.

Hardware requirement: Llama-3.1-8B-Instruct requires N300 or higher for reliable operation. It consistently exhausts DRAM on N150 and P300c. Qwen3-0.6B is the recommended choice for those devices.

If you've accepted Meta's license terms and are running on N300/T3K/P100/Galaxy, you can download Llama-3.1-8B-Instruct:

mkdir -p ~/models && hf download meta-llama/Llama-3.1-8B-Instruct \
  --local-dir ~/models/Llama-3.1-8B-Instruct

What gets downloaded (~16GB):


Next Steps

You've successfully downloaded your model and are ready to run inference.

Next: Verify Your Setup → verify-installation

Learn More