← Back to Quantization

mixed int5/int6 with QAT

Quantization
Used in
1 PRs
Best BPB
1.4222
Avg BPB
1.4222

Hyperparameters Across PRs

pr_numberbitsscope
619int5 for MLP weights, int6 for attention/bigram-sensitive weights