← All Models
|
Gemma - Google's lightweight open models built from Gemini research:
gemma-2b
gemma-7b
gemma-2-2b
gemma-2-9b
gemma-2-27b
gemma-3-270m
gemma-3-1b-pt
gemma-3-4b-pt
gemma-3-12b-pt
gemma-3-27b-pt
google/gemma-3-4b-pt
📊 Model Parameters
Total Parameters
4,971,331,952
Context Length
2,048
Hidden Size
0
Layers
0
Attention Heads
0
KV Heads
0
💾 Memory Requirements
FP32 (Full)
18.52 GB
FP16 (Half)
9.26 GB
INT8 (Quantized)
4.63 GB
INT4 (Quantized)
2.31 GB
🔑 KV Cache (Inference)
Per Token (FP16)
0 B
Max Context FP32
0.0 MB
Max Context FP16
0.0 MB
Max Context INT8
0.0 MB
⚙️ Model Configuration
Attention Configuration
Tied Embeddings
Yes
Special Tokens
BOS Token ID
Not set
Pad Token ID
Not set
EOS Token ID
Not set
Data Type
Model Dtype
bfloat16
Layer Types:
Attention
MLP/FFN
Normalization
Embedding
Attention
MLP
Norm
Embedding
Clear
Expand All
Collapse All