GPUs with 16–undefined GB VRAM
Browse 38 GPUs with 16–undefined GB VRAM compatible with running LLM models locally. Compare VRAM, memory bandwidth, and AI performance.
← Show all GPUsWhich GPU Do You Need for AI?
The amount of VRAM is the most important specification for running LLMs locally. Most 7B parameter models require 4–8 GB of VRAM at common quantization levels, while 70B models need 24–48 GB. Memory bandwidth determines how fast the model generates tokens — faster bandwidth means faster responses.
GPU List
AMD Instinct MI210
AMD · CDNA 2
AMD Instinct MI250X
AMD · CDNA 2
AMD Instinct MI300X
AMD · CDNA 3
AMD Radeon PRO W7800
AMD · RDNA 3
AMD Radeon PRO W7900
AMD · RDNA 3
AMD Radeon RX 6800
AMD · RDNA 2
AMD Radeon RX 6800 XT
AMD · RDNA 2
AMD Radeon RX 6900 XT
AMD · RDNA 2
AMD Radeon RX 7800 XT
AMD · RDNA 3
AMD Radeon RX 7900 XT
AMD · RDNA 3
AMD Radeon RX 7900 XTX
AMD · RDNA 3
Intel Arc A770 16GB
Intel · Alchemist
NVIDIA A100 40GB PCIe
NVIDIA · Ampere
NVIDIA A100 80GB SXM
NVIDIA · Ampere
NVIDIA A40
NVIDIA · Ampere
NVIDIA GeForce RTX 3090
NVIDIA · Ampere
NVIDIA GeForce RTX 3090 Ti
NVIDIA · Ampere
NVIDIA GeForce RTX 4060 Ti 16GB
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4070 Ti SUPER
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4080
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4080 SUPER
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4090
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 5070 Ti
NVIDIA · Blackwell
NVIDIA GeForce RTX 5080
NVIDIA · Blackwell
NVIDIA GeForce RTX 5090
NVIDIA · Blackwell
NVIDIA H100 PCIe
NVIDIA · Hopper
NVIDIA H100 SXM
NVIDIA · Hopper
NVIDIA L4
NVIDIA · Ada Lovelace
NVIDIA L40
NVIDIA · Ada Lovelace
NVIDIA L40S
NVIDIA · Ada Lovelace