← All Models
|
Other Chinese/Asian AI Labs:
Seed-OSS-36B-Base
Seed-OSS-36B-Base-woSyn
Kimi-K2-Thinking
LongCat-Flash-Chat
GLM-4.5
GLM-4.5-Air
GLM-4.6
GLM-4.6V
GLM-4.6V-Flash
dots.llm1.base
zai-org/GLM-4.6V-Flash
📊 Model Parameters
Total Parameters
9,672,020,480
Context Length
2,048
Hidden Size
0
Layers
0
Attention Heads
0
KV Heads
0
💾 Memory Requirements
FP32 (Full)
36.03 GB
FP16 (Half)
18.02 GB
INT8 (Quantized)
9.01 GB
INT4 (Quantized)
4.50 GB
🔑 KV Cache (Inference)
Per Token (FP16)
0 B
Max Context FP32
0.0 MB
Max Context FP16
0.0 MB
Max Context INT8
0.0 MB
⚙️ Model Configuration
Attention Configuration
Tied Embeddings
No
Special Tokens
BOS Token ID
Not set
Pad Token ID
Not set
EOS Token ID
Not set
Data Type
Model Dtype
Not set
Layer Types:
Attention
MLP/FFN
Normalization
Embedding
Attention
MLP
Norm
Embedding
Clear
Expand All
Collapse All