NVIDIAOrinAI Box

Best AI Models for NVIDIA Jetson AGX Orin 32GB

Memory:32 GB Unified·Bandwidth:204.8 GB/s·GPU Cores:1792 GPU cores·CPU Cores:8 CPU cores·Neural Engine:200.0 TOPS

32 GB unified − 3.5 GB OS overhead = 28.5 GB available for AI models

32 GB positions this hardware in the professional tier for local AI. Most popular open-source models run comfortably, and even large 70B parameter models are accessible at lower quantization levels.

This memory amount is a sweet spot for enthusiasts and professionals. You can run 13B–30B models like DeepSeek R1 Distill at Q5 or Q6 quality with smooth token generation, and 7B models at near-lossless precision. The 70B class of models (Llama 3 70B, Qwen 72B) becomes possible at Q2–Q3 quantization, though with some quality trade-off. For day-to-day use with coding assistants, chat models, and reasoning tasks, this tier delivers an excellent experience.

Runs Well

  • 7B–13B models at Q6–Q8 quality
  • 14B–30B models at Q4–Q5 quality
  • Small models (3B–7B) at FP16 precision
  • Vision-language models at good quality

Challenging

  • 70B models only at Q2–Q3 (noticeable quality loss)
  • Large context windows with 30B+ models

What LLMs Can NVIDIA Jetson AGX Orin 32GB Run?

29 models · 7 good

Showing compatibility for NVIDIA Jetson AGX Orin 32GB

LLM models compatible with NVIDIA Jetson AGX Orin 32GB — ranked by performance
ModelVRAMGrade
Q4_K_M·6.5 t/s tok/s·131K ctx·GOOD FIT
20.5 GBA81
Q4_K_M·6.5 t/s tok/s·33K ctx·GOOD FIT
20.5 GBA81
Qwen3 32B32B
Q4_K_M·6.7 t/s tok/s·41K ctx·GOOD FIT
19.8 GBA78
Q4_K_M·6.2 t/s tok/s·4K ctx·GOOD FIT
21.4 GBA84
QwQ 32B32B
Q4_K_M·6.6 t/s tok/s·41K ctx·GOOD FIT
20.0 GBA80
Gemma 3 27B IT27.4B
Q4_K_M·7.4 t/s tok/s·131K ctx·GOOD FIT
18.1 GBA72
Q4_K_M·7.4 t/s tok/s·8K ctx·GOOD FIT
18.0 GBA71
GPT OSS 20B21.5B
Q4_K_M·10.0 t/s tok/s·131K ctx·FAIR FIT
13.3 GBB57
Q4_K_M·8.8 t/s tok/s·33K ctx·FAIR FIT
15.1 GBB62
Q4_K_M·4.7 t/s tok/s·33K ctx·FAIR FIT
28.6 GBB56
Qwen3 4B4B
Q4_K_M·46.1 t/s tok/s·41K ctx·EASY RUN
2.9 GBC30
Q4_K_M·67.2 t/s tok/s·131K ctx·EASY RUN
2.0 GBD28
Phi 22.8B
Q4_K_M·50.4 t/s tok/s·2K ctx·EASY RUN
2.6 GBD29
Q4_K_M·131.8 t/s tok/s·2K ctx·EASY RUN
1.0 GBD27
Q4_K_M·201.7 t/s tok/s·131K ctx·EASY RUN
0.7 GBD26
Q4_K_M·201.7 t/s tok/s·33K ctx·EASY RUN
0.7 GBD26

NVIDIA Jetson AGX Orin 32GB Specifications

Brand
NVIDIA
Chip
Orin
Type
AI Box
Unified Memory
32 GB
Memory Bandwidth
204.8 GB/s
GPU Cores
1792
CPU Cores
8
Neural Engine
200.0 TOPS
Release Date
2022-03-22

Get Started

Ollama (Recommended)

$curl -fsSL https://ollama.com/install.sh | sh
$ollama run llama3:8b

LM Studio

LM Studio

Download LM Studio, search for a model, and run it with one click.

Devices to Consider

Similar devices and upgrades with more memory or higher bandwidth

Frequently Asked Questions

Can NVIDIA Jetson AGX Orin 32GB run DeepSeek R1 Distill Qwen 32B?

Yes, the NVIDIA Jetson AGX Orin 32GB with 32 GB unified memory can run DeepSeek R1 Distill Qwen 32B, Qwen2.5 Coder 32B Instruct, Qwen3 32B, and 1158 other models. 21 models achieve excellent performance, and 237 run at good quality. Apple Silicon's unified memory architecture lets the GPU access the full memory pool without copying data, making it efficient for AI workloads.

How much memory is available for AI on NVIDIA Jetson AGX Orin 32GB?

The NVIDIA Jetson AGX Orin 32GB has 32 GB unified memory. After macOS reserves ~3.5 GB for the operating system, approximately 28.5 GB is available for AI models. Unlike discrete GPUs where VRAM is separate from system RAM, Apple Silicon shares one memory pool between the CPU and GPU — this means no data copying overhead, but you share memory with macOS and open apps.

Is NVIDIA Jetson AGX Orin 32GB good for AI?

With 32 GB unified memory and 204.8 GB/s bandwidth, the NVIDIA Jetson AGX Orin 32GB is very good for running local AI models. It supports 258 models at good quality or better. This is a strong configuration for AI — 7B models run at maximum quality, and you can comfortably handle 14B models like DeepSeek R1 Distill 14B and larger.

What's the best model for NVIDIA Jetson AGX Orin 32GB?

The top-rated models for the NVIDIA Jetson AGX Orin 32GB are DeepSeek R1 Distill Qwen 32B, Qwen2.5 Coder 32B Instruct, Qwen3 32B. For general chat, instruction-tuned 7B models give the best speed-to-quality ratio. For coding or reasoning, a 14B model at Q4_K_M is a sweet spot.

How fast is NVIDIA Jetson AGX Orin 32GB for AI inference?

With 204.8 GB/s memory bandwidth, the NVIDIA Jetson AGX Orin 32GB achieves approximately 32 tok/s on a 7B model at Q4_K_M — that's comfortable for real-time chat. A 14B model runs at ~16 tok/s. Apple Silicon achieves high efficiency (~70%) thanks to unified memory — there's no PCIe bottleneck between CPU and GPU.

tok/s = (204.8 GB/s ÷ model GB) × efficiency

Apple Silicon achieves ~70% bandwidth efficiency thanks to unified memory and Metal acceleration.

Estimated speed on NVIDIA Jetson AGX Orin 32GB

Real-world results typically within ±20%.

Learn more about tok/s estimation →

Can I run AI offline on NVIDIA Jetson AGX Orin 32GB?

Yes — once you download a model, it runs entirely on the NVIDIA Jetson AGX Orin 32GB without internet. Applications like Ollama and LM Studio make it straightforward to download, manage, and run models locally. All your conversations stay private on your device with zero data sent to external servers. This is one of the key advantages of local AI: complete privacy, no API costs, and no rate limits.