Key checks
- 12 Million Token Context Window: The company's official launch post and website both explicitly claim a 12 million token context window, describing it as a 'research result' that enables processing entire codebases in one pass.
- Sub-Quadratic Sparse-Attention Architecture (SSA): Documentation confirms SubQ is built on a ground-up redesign of attention called SSA, which allows compute to grow linearly rather than quadratically with context length.
- Speed and Cost Efficiency: Benchmarks provided by the company and cited in funding news state the model is 52x faster than FlashAttention at 1M tokens and operates at roughly 20% (1/5) the cost of leading competitors.