mistralai/Voxtral-Small-24B-2507

📊 Model Parameters

Total Parameters 24,261,800,960
Context Length 2,048
Hidden Size 5120
Layers 0
Attention Heads 0
KV Heads 0

💾 Memory Requirements

FP32 (Full) 90.38 GB
FP16 (Half) 45.19 GB
INT8 (Quantized) 22.60 GB
INT4 (Quantized) 11.30 GB

🔑 KV Cache (Inference)

Per Token (FP16) 0 B
Max Context FP32 0.0 MB
Max Context FP16 0.0 MB
Max Context INT8 0.0 MB

⚙️ Model Configuration

Core Architecture

Vocabulary Size131,072
Hidden Size5,120

Attention Configuration

Tied EmbeddingsYes

Special Tokens

BOS Token IDNot set
Pad Token IDNot set
EOS Token IDNot set

Data Type

Model Dtypebfloat16
Layer Types:
Attention
MLP/FFN
Normalization
Embedding