PR #1461
openRecord Submission: HDC_1_Step_Grad_DSV_Radial_Slyvester_Hadamard_Matrix_Symmetry_Language_Model_val_bpb: 0.4118
by viasky657View on GitHub
val_bpb
0.4118
Architecture
Hybrid
Optimizer
—
Artifact Size
15,943,520 bytes
Training Techniques
Architecture
BigramHash
Hash-based bucketed next-token prediction pipeline with fingerprint routing and bucket frequency tables.
parameters: {"table_bits":19,"embed_dim":16}
DirectionalSemanticVec
Directional semantic vector layer using forward/backward semantic bundles and skip-bigram lags as the primary predictive signal.
parameters: {"lags":[2,3,4,5]}
Hadamard Matrix
Uses a Sylvester Hadamard matrix in the learning / codebook pipeline.
parameters: null
weight tying
Codebook / embedding-style shared representation is implied by the compact factorized pipeline.
parameters: null
Compression
lzma
level: 9
Regularization
magnitude pruning
parameters: {"threshold":1}
Other
other
1-step random gradient learning / single-iteration NMF-style update used to fit the hash-gradient factors.
parameters: {"nmf_max_iter":1}
Evaluation
multi-seed evaluation
parameters: {"seeds":[42,7,1337],"runs":3}
Sequence Length
sequence_length
train_length: null
eval_length: null
Novel Contributions
- Directional Semantic Vector (DSV) fallback as the primary predictive mechanism
- 1-step hash-gradient / NMF factorization pipeline
- Sylvester Hadamard matrix and codebook-based learning
- Fingerprint-routed bucket prediction with collision fallback
- Skip-bigram semantic lags 2–5
- Multi-seed merged training across three seeds
- Artifact compression to fit within the 16 MB limit