All GPUs for Running LLMs Locally
Browse 56 GPUs compatible with running LLM models locally. Compare VRAM, memory bandwidth, and AI performance.
Which GPU Do You Need for AI?
The amount of VRAM is the most important specification for running LLMs locally. Most 7B parameter models require 4–8 GB of VRAM at common quantization levels, while 70B models need 24–48 GB. Memory bandwidth determines how fast the model generates tokens — faster bandwidth means faster responses.
GPU List
NVIDIA GeForce RTX 4060 Ti 16GB
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4060 Ti 8GB
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4070
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4070 SUPER
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4070 Ti
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4070 Ti SUPER
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4080
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4080 SUPER
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 4090
NVIDIA · Ada Lovelace
NVIDIA GeForce RTX 5070
NVIDIA · Blackwell
NVIDIA GeForce RTX 5070 Ti
NVIDIA · Blackwell
NVIDIA GeForce RTX 5080
NVIDIA · Blackwell
NVIDIA GeForce RTX 5090
NVIDIA · Blackwell
NVIDIA H100 PCIe
NVIDIA · Hopper
NVIDIA H100 SXM
NVIDIA · Hopper
NVIDIA L4
NVIDIA · Ada Lovelace
NVIDIA L40
NVIDIA · Ada Lovelace
NVIDIA L40S
NVIDIA · Ada Lovelace
NVIDIA RTX 4000 Ada Generation
NVIDIA · Ada Lovelace
NVIDIA RTX 5000 Ada Generation
NVIDIA · Ada Lovelace
NVIDIA RTX 6000 Ada Generation
NVIDIA · Ada Lovelace
NVIDIA RTX A4000
NVIDIA · Ampere
NVIDIA RTX A5000
NVIDIA · Ampere
NVIDIA RTX A6000
NVIDIA · Ampere
NVIDIA T4
NVIDIA · Turing
NVIDIA V100 SXM2 32GB
NVIDIA · Volta