← Back to Quantization
mixed int6
QuantizationUsed in
19 PRs
Best BPB
0.0972
Avg BPB
1.0655
Submissions
PR #135by unnir
1.1539PR #174by Julz19
1.1537PR #339by sheeki03
1.1364PR #398by felipe-parodi
1.1213PR #581by teddyoweh
1.0698PR #649by pall23-mech
1.2073PR #684by DeepReinforce
1.0574PR #698by hesong0222-dev
1.1642PR #811by quietsmile
0.4377PR #922by greqone
0.0972PR #993by aerosta
0.9631PR #1052by demouo
1.1978PR #1427by kjahan
1.2092PR #1438by sabdulmajid
1.2029PR #1465by sisegod
1.1381PR #1569by abbudjoe
1.3576PR #1664by zoharb157
1.0980PR #1665by mrbese
1.3571PR #1696by kings-crown
1.1224Hyperparameters Across PRs
| pr_number | bits | scope |
|---|---|---|
| 135 | 6 | MLP and attention weight matrices; FP16 passthrough for tied embeddings and last 2 layers' Key projections |
| 174 | 6 | large MLP and attention matrices |
| 339 | 6 | model weights |
| 398 | 6 | all |
| 581 | 6 | model weights |
| 649 | 6 | all |
| 684 | 6 | model weights |
| 698 | 6 | all |
| 811 | 6 | model weights |
| 922 | 6 | model |
| 993 | 6 | post-training mixed |
| 1052 | 6 | artifact |
| 1427 | 6 | model weights |
| 1438 | 6 | mlp;attn;embed |
| 1465 | 6 | embeddings |
| 1569 | 6 | default export |
| 1664 | 6 | all |
| 1665 | 6 | MLP, attention, and Mamba projection weights |
| 1696 | 6 | attention/MLP banks |