All GPUs for Running LLMs Locally
Browse 56 GPUs compatible with running LLM models locally. Compare VRAM, memory bandwidth, and AI performance.
Which GPU Do You Need for AI?
The amount of VRAM is the most important specification for running LLMs locally. Most 7B parameter models require 4–8 GB of VRAM at common quantization levels, while 70B models need 24–48 GB. Memory bandwidth determines how fast the model generates tokens — faster bandwidth means faster responses.
GPU List
AMD Instinct MI210
AMD · CDNA 2
AMD Instinct MI250X
AMD · CDNA 2
AMD Instinct MI300X
AMD · CDNA 3
AMD Radeon PRO W7800
AMD · RDNA 3
AMD Radeon PRO W7900
AMD · RDNA 3
AMD Radeon RX 6700 XT
AMD · RDNA 2
AMD Radeon RX 6800
AMD · RDNA 2
AMD Radeon RX 6800 XT
AMD · RDNA 2
AMD Radeon RX 6900 XT
AMD · RDNA 2
AMD Radeon RX 7600
AMD · RDNA 3
AMD Radeon RX 7700 XT
AMD · RDNA 3
AMD Radeon RX 7800 XT
AMD · RDNA 3
AMD Radeon RX 7900 XT
AMD · RDNA 3
AMD Radeon RX 7900 XTX
AMD · RDNA 3
Intel Arc A750
Intel · Alchemist
Intel Arc A770 16GB
Intel · Alchemist
Intel Arc B580
Intel · Battlemage
NVIDIA A100 40GB PCIe
NVIDIA · Ampere
NVIDIA A100 80GB SXM
NVIDIA · Ampere
NVIDIA A40
NVIDIA · Ampere
NVIDIA GeForce GTX 1080 Ti
NVIDIA · Pascal
NVIDIA GeForce RTX 3060 12GB
NVIDIA · Ampere
NVIDIA GeForce RTX 3060 Ti
NVIDIA · Ampere
NVIDIA GeForce RTX 3070
NVIDIA · Ampere
NVIDIA GeForce RTX 3070 Ti
NVIDIA · Ampere
NVIDIA GeForce RTX 3080
NVIDIA · Ampere
NVIDIA GeForce RTX 3080 Ti
NVIDIA · Ampere
NVIDIA GeForce RTX 3090
NVIDIA · Ampere
NVIDIA GeForce RTX 3090 Ti
NVIDIA · Ampere
NVIDIA GeForce RTX 4060
NVIDIA · Ada Lovelace