LLM Inference Memory Calculator
English
Model Parameters (B)
Select example model
Computation Precision
GPU Model
NVIDIA A100 (80GB)
Model Weight Memory:
14
GB
Inference Extra Memory (≈20%):
2.8
GB
Total Memory Required:
16.8
GB
Required GPU Count:
1
units NVIDIA A100 (80GB)