Collected molecules will appear here. Add from search or explore.
A long-range genomic foundation model utilizing the sub-quadratic Hyena operator to process DNA sequences up to 1 million tokens in length.
stars
776
forks
105
HyenaDNA is a high-impact project from HazyResearch (Stanford), a lab known for pioneering sub-quadratic architectures like S4 and Monarch. Its defensibility stems from the extreme computational cost and specialized domain knowledge required to train genomic foundation models at 1M+ context lengths, which traditional transformers (like DNABERT) cannot achieve due to quadratic scaling. With 770+ stars and significant fork activity, it has established itself as a research standard in the 'Long Context' genomics niche. While frontier labs like OpenAI/Anthropic are unlikely to target genomics directly, Google/DeepMind (via AlphaFold and Enformer) represent a medium platform risk. The primary technical threat comes from more recent State Space Models (SSMs) like Mamba, which have shown superior performance-per-compute over Hyena in some benchmarks. However, the ecosystem lock-in and the specific pre-trained weights for the human genome provide a significant moat for researchers in computational biology. The 0.0 velocity likely indicates the project has reached maturity or that the team's focus has shifted to the 'Evo' project (a successor from the same lab), but the codebase remains a critical reference for non-attention-based sequence modeling.
TECH STACK
INTEGRATION
library_import
READINESS