PR #1238
openNon-record: TurboQuant mixed-precision int4/int5 (val_bpb=1.1521)
by ibarrajoView on GitHub
val_bpb
1.1521
Architecture
Transformer
Optimizer
—
Artifact Size
13.4 MB
Training Techniques
Quantization
mixed int4/int5
bits: null
scope: Q/K int5, V/O and MLP int4 in middle layers; boundary layers int5
Test-Time Training
score-first TTT
parameters: null
Novel Contributions
- Role-based mixed-precision weight quantization using TurboQuant-guided layer sensitivity
- Keeping Q/K projections at int5 while quantizing V/O and MLP weights to int4 in middle layers
- Using int5 for boundary layers to preserve quality
- Negative result showing int3 weight quantization is unusable for this model
- Observation that weight quantization sensitivity differs from KV cache activation sensitivity