LLM VRAM Calculator

Estimate GPU memory requirements for LLM training and inference

Quick Model Presets

Hardware

GPU Model

Single GPU memory estimation

Model

Mode
Precision

BF16 recommended for training on modern GPUs (Ampere+). FP16 for older GPUs (V100, T4).

Runtime

1 1024
128 131072

Training Options

Optimizer

Keep FP32 copy of weights for stability. Recommended for FP16, optional for BF16.

4 256

Advanced Options