Palo Alto startup Zyphra on Tuesday released ZAYA1-8B, an 8-billion-parameter reasoning model trained entirely on AMD Instinct MI300 GPUs, demonstrating competitive performance against GPT-5-High and DeepSeek-V3.2 while using only 760 million active parameters. According to Zyphra’s announcement, the model is available under Apache 2.0 license and marks a significant validation of AMD’s alternative to NVIDIA’s dominant position in AI training infrastructure.
Meanwhile, Miami-based Subquadratic emerged from stealth claiming its SubQ 1M-Preview model achieves 1,000x efficiency gains through fully subquadratic architecture, though AI researchers are demanding independent verification of the extraordinary claims.
ZAYA1-8B Architecture and Training Breakthrough
ZYAYA1-8B employs a mixture-of-experts (MoE) architecture that activates only 760 million of its 8 billion total parameters during inference, creating what Zyphra calls “intelligence density.” The model’s training on AMD Instinct MI300 GPUs represents the first major validation that AMD’s platform can produce competitive AI models without NVIDIA hardware.
VentureBeat reported that Zyphra used a “full-stack innovation” approach spanning architecture, training techniques, and hardware optimization. The company’s previous Zamba model, released in 2024, mimicked cortex-hippocampus interactions to share information across sequential layers, providing the foundation for ZAYA1-8B’s efficiency gains.
The model is available on Hugging Face and can be tested through Zyphra Cloud, the company’s inference platform. Enterprise developers can begin customization immediately under the permissive Apache 2.0 license.
SubQ’s Extraordinary Efficiency Claims Draw Skepticism
Subquadratic’s SubQ 1M-Preview claims to be the first large language model built on fully subquadratic architecture, where compute grows linearly rather than quadratically with context length. At 12 million tokens, the company says its architecture reduces attention compute by nearly 1,000 times compared to frontier models.
The startup raised $29 million in seed funding from investors including Tinder co-founder Justin Mateen and former SoftBank Vision Fund partner Javier Villamizar, with The New Stack reporting a $500 million valuation. Subquadratic is launching three products in private beta: an API, SubQ Code command-line agent, and SubQ Search tool.
However, AI researchers have responded with skepticism to the claims. AI engineer Will Depue noted concerns about the benchmarking methodology, while developer Stepan Goncharov called the results “very interesting cherry-picked benchmarks.” The research community is demanding independent verification before accepting the 1,000x efficiency claims.
Google Advances Gemma 4 with Multi-Token Prediction
Google released Multi-Token Prediction (MTP) drafters for its Gemma 4 model family, achieving up to 3x speedup through speculative decoding without quality degradation. According to Google’s blog post, the technique addresses memory-bandwidth bottlenecks in standard LLM inference.
The MTP approach pairs heavy target models like Gemma 4 31B with lightweight drafters that predict multiple future tokens simultaneously. While the target model processes one token, the drafter utilizes idle compute to generate several candidate tokens, which are then verified by the target model.
Google tested the speedup across multiple frameworks including LiteRT-LM, MLX, Hugging Face Transformers, and vLLM. The technique particularly benefits consumer-grade hardware where processors spend most time moving parameters from VRAM to compute units rather than performing actual calculations.
Timer-XL Extends Transformer Architecture to Time Series
Researchers at Tsinghua University’s THUML lab released Timer-XL, a decoder-only Transformer foundation model for time series forecasting that handles variable input and output lengths in a single unified model. According to research published on Towards Data Science, the model introduces TimeAttention, a specialized attention mechanism for long-context forecasting.
Timer-XL supports non-stationary univariate series, complex multivariate dynamics, and covariate-informed contexts with exogenous variables. Unlike models such as Tiny-Time-Mixers that require different versions for different input or output lengths, Timer-XL uses a single architecture for all cases.
The THUML lab previously released milestone models including iTransformer, TimesNet, and the original Timer model. Timer-XL can be trained from scratch or pretrained on large datasets, with optional fine-tuning for improved performance on specific use cases.
Training Efficiency and Hardware Diversification Trends
The releases highlight two key trends in AI development: the pursuit of training efficiency and hardware diversification beyond NVIDIA’s ecosystem. ZAYA1-8B’s success on AMD hardware demonstrates that competitive models can be trained without NVIDIA GPUs, potentially reducing costs and supply chain dependencies for AI companies.
Zyphra’s mixture-of-experts architecture achieves competitive performance with 95% fewer active parameters than trillion-parameter models from major labs. This efficiency focus aligns with broader industry interest in smaller, more practical models that can run on consumer hardware while maintaining strong capabilities.
Google’s MTP drafters for Gemma 4 address the fundamental memory-bandwidth bottleneck that limits inference speed across all current architectures. The 3x speedup without quality loss represents a significant practical improvement for developers deploying models in production environments.
What This Means
These architectural advances signal a maturation of AI model development beyond simply scaling parameters. ZAYA1-8B proves that AMD hardware can compete with NVIDIA for AI training, potentially breaking the current hardware monopoly and reducing costs. The model’s efficiency demonstrates that careful architecture design can achieve competitive results with dramatically fewer resources.
Subquadratic’s claims, if verified, would represent a fundamental breakthrough in how AI systems scale with context length. However, the research community’s skepticism reflects the need for rigorous independent validation of extraordinary efficiency claims. The startup’s $500 million valuation based on unverified claims highlights both the potential impact and risk of architectural breakthroughs.
Google’s MTP approach offers immediate practical benefits for existing Transformer architectures, providing a clear path to faster inference without requiring fundamental model redesigns. This incremental but significant improvement demonstrates how optimization techniques can extend the capabilities of current architectures while researchers work on next-generation designs.
FAQ
How does ZAYA1-8B achieve competitive performance with only 760 million active parameters?
ZYAYA1-8B uses a mixture-of-experts (MoE) architecture that selectively activates only 760 million of its 8 billion total parameters for each inference task. This approach, combined with architectural innovations inspired by cortex-hippocampus interactions, allows the model to maintain performance while using significantly less compute than traditional dense models.
What makes Subquadratic’s efficiency claims so controversial?
Subquadratic claims its SubQ model reduces attention compute by 1,000x compared to other frontier models through fully subquadratic architecture. AI researchers are skeptical because such gains would represent a fundamental breakthrough in how models scale with context length, yet the company has not provided independent verification or detailed technical papers explaining how this is achieved.
How does Google’s Multi-Token Prediction improve inference speed?
MTP uses a lightweight “drafter” model to predict multiple future tokens while the main model processes a single token. The main model then verifies these predictions in parallel, effectively utilizing idle compute resources. This speculative decoding approach can achieve up to 3x speedup without any loss in output quality or reasoning capability.






