Field notes
There is free lunch: Why linear attention is necessary for context scaling
Giving insights as to why linear attention and mamba may be necessary for expressive context scaling.
Coming soon
Research, engineering write-ups, and field notes from the team. Proper posts are on the way. here’s what’s coming.
Giving insights as to why linear attention and mamba may be necessary for expressive context scaling.
An overview of efficient circuit discovery for post-SAE mechanistic interpretability for high speed knowledge and safety distillation.
A behind-the-scenes look into our work, and the niche research we think view as promising for the future of small but efficient language models
Want the posts as they land? Start an account and we’ll let you know.
Get Early Access