Add Hybrid Depth-Recurrent Transformer submission#341
Add Hybrid Depth-Recurrent Transformer submission#341tobiascanavesi wants to merge 1 commit intoopenai:mainfrom
Conversation
Community Review — Add Hybrid Depth-Recurrent Transformer submissionBPB: 0.004 (cache parse — may be delta/std, not val_bpb; check PR title) | Compliance: LOOKS CLEAN — pure-neural submission, no TTT/SLOT/n-gram-cache What I found in the code (head SHA Static code review found no TTT adaptation function, no SLOT optimization loop, no n-gram-cache class, and no pre-quant val-token fine-tune. The eval path uses the standard sliding-window stride-64 pattern. The submission is a pure-neural architecture iteration on the standard SP1024/SP4096/SP8192 baseline. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=768, layers=22, vocab=1024, code=58154 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending the usual record-track checks (3-seed validation, under-16MB artifact cap, ≤600s train + ≤600s eval on 8×H100 SXM). No compliance flags from the classification pass — this looks like a clean pure-neural iteration on the standard baseline. Auto-classification caveat: this review was drafted by the AST-based classifier. If there's a non-standard eval mechanism (logit postprocessing, hedge mixing, etc.) that I missed because it's factored into a helper file or a non-standard function name, please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=768, layers=22, vocab=1024, code=58154 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Hybrid Depth-Recurrent Transformer
Testing this new architecture that solves the int8 quantization compounding problem in depth-recurrent transformers.
Key Insight
Standard depth-recurrence shares all weights across loop iterationsm int8 rounding errors compound on every loop (0.40 BPB gap). The hybrid keeps precision-sensitive layers near input/output as unique weights, while only the bulk middle layers are shared and looped.
Result: quantization gap reduced from 0.40 to near-zero (-0.004 BPB).
Architecture
Techniques
Preliminary Results (2×H100)
8×H100 run pending, expecting significant improvement with full compute.
Reproduce