You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
“A clean, from-scratch implementation of the OLMo architecture with KV caching, RoPE, and an efficient autoregressive inference pipeline. Designed as a minimal yet extensible foundation for post-training research, including RLHF, preference optimization, and reasoning-focused systems.”
OLMo-7B QLoRA fine-tuned on ESOL solubility regression. RDKit-Augmented Prompts + Regression Head on 4096-dim hidden states. Best RMSE: 0.8582. DeepChem GSoC 2026.
Empirical analysis quantifying the 0.009 ROC-AUC gap between PyTorch GNNs (2D graphs) and OLMo-7B (1D SMILES) on Tox21. Exposes the Random Split illusion and provides the baseline justification for Topological State Machine (TSM) integration.