DeepSeek released its V4 model on Monday, delivering near-GPT-5.5 performance at one-sixth the cost, while U.S. startup Poolside launched its Laguna XS.2 coding model — marking a significant week for open source AI competition against proprietary giants. According to DeepSeek’s announcement, the 1.6-trillion-parameter model is available under MIT license through Hugging Face and their API.
The releases come as open source models increasingly challenge closed-source leaders from OpenAI and Anthropic, with Chinese companies leading cost-efficient alternatives and American startups targeting specialized use cases.
DeepSeek V4 Matches Frontier Performance at Fraction of Cost
DeepSeek V4 approaches the performance of Claude Opus 4.7 and GPT-5.5 while offering API pricing approximately 83% lower than proprietary alternatives. DeepSeek AI researcher Deli Chen described the release as a “labor of love” developed over 484 days since V3’s launch.
The Mixture-of-Experts architecture enables efficient inference while maintaining competitive capabilities across reasoning, coding, and general knowledge tasks. According to VentureBeat’s analysis, the model represents a “second DeepSeek moment” following the company’s January breakthrough with R1.
The model is immediately available through DeepSeek’s API and on Hugging Face for local deployment. Enterprise users can modify and deploy the model commercially without licensing restrictions.
Poolside Launches Specialized Coding Models from Silicon Valley
Poolside, a San Francisco-based startup founded in 2023, announced two Laguna models optimized for agentic coding workflows. The company positioned itself as a U.S. alternative to Chinese open source offerings, targeting government and enterprise customers requiring domestic AI solutions.
The Laguna XS.2 model focuses on autonomous coding tasks, including writing code, using third-party tools, and taking actions without human intervention. Poolside also launched “pool,” a coding agent harness, and “shimmer,” a web-based development environment optimized for mobile devices.
Poolside post-training engineer George Grigorev told VentureBeat that government agencies might prefer Poolside over proprietary U.S. labs due to open source transparency and local deployment capabilities. The models are available for immediate download and commercial use.
https://x.com/eisokant/status/2049142230397370537
Xiaomi Advances Agentic AI with MiMo V2.5 Series
Xiaomi released MiMo-V2.5 and MiMo-V2.5-Pro under MIT license, targeting agentic “claw” tasks where AI systems complete autonomous actions through third-party messaging platforms. According to Xiaomi’s announcement, both models rank among the most token-efficient options for enterprise automation.
The Pro model achieved 63.8% performance on ClawEval benchmarks while maintaining low token consumption — critical as services like GitHub Copilot move to usage-based billing. The models excel at marketing content creation, account management, email organization, and scheduling tasks.
Both models are available through Hugging Face for immediate enterprise deployment. The token efficiency positions them as cost-effective alternatives for businesses implementing AI agents at scale.
Infrastructure Innovation Accelerates Open Source Adoption
Runpod launched Flash, an open source Python tool eliminating Docker containerization requirements for AI development. The MIT-licensed platform enables faster iteration cycles and reduces deployment complexity for serverless GPU infrastructure.
Runpod CTO Brennen Smith told VentureBeat the tool makes it “as easy as possible to bring together the cosmos of different AI tooling in a function call.” Flash supports polyglot pipelines, routing preprocessing to CPU workers before transferring to high-end GPUs for inference.
The platform integrates with coding assistants like Claude Code, Cursor, and Cline, enabling autonomous hardware orchestration with minimal friction. Features include low-latency APIs, queue-based batch processing, and persistent multi-datacenter storage for production deployments.
Enterprise Multimodal Capabilities Expand
New research demonstrates multimodal responses without requiring multimodal embeddings through Proxy-Pointer RAG architecture. The approach treats documents as hierarchical semantic trees rather than text chunks, enabling image and table retrieval alongside text responses.
The technique addresses enterprise chatbot limitations in returning relevant images from source documents. Use cases span real estate property images, technical maintenance diagrams, and equipment specifications where visual context enhances text-only responses.
The open source implementation maintains scalability while minimizing computational costs compared to traditional multimodal embedding approaches. The architecture supports diverse document types including brochures, manuals, and technical specifications.
What This Means
Open source AI models are rapidly closing the performance gap with proprietary alternatives while offering significant cost advantages and deployment flexibility. DeepSeek V4’s pricing at one-sixth the cost of leading models demonstrates how open source can disrupt enterprise AI economics.
The geographic distribution of innovation — Chinese companies focusing on efficiency, U.S. startups targeting specialized applications — suggests a maturing ecosystem with distinct competitive advantages. Enterprise adoption will likely accelerate as licensing clarity and infrastructure tooling improve.
Specialized models like Poolside’s coding-focused Laguna and Xiaomi’s agentic MiMo series indicate market segmentation toward task-specific optimization rather than general-purpose scaling. This trend favors organizations with clear use cases over those seeking generic AI capabilities.
FAQ
How does DeepSeek V4 achieve similar performance at lower cost?
DeepSeek V4 uses Mixture-of-Experts architecture that activates only relevant model components for each task, reducing computational requirements. The open source licensing eliminates proprietary markup, while efficient training techniques lower operational costs.
Can enterprises use these open source models commercially?
Yes, all mentioned models (DeepSeek V4, Poolside Laguna, Xiaomi MiMo) are released under MIT license, permitting commercial modification, deployment, and redistribution without royalties or usage restrictions.
What advantages do specialized models offer over general-purpose AI?
Specialized models like Poolside’s coding focus and Xiaomi’s agentic optimization deliver superior performance on specific tasks while requiring fewer computational resources. They often integrate better with existing workflows and offer more predictable costs for targeted use cases.
Related news
Sources
- American AI startup Poolside launches free, high-performing open model Laguna XS.2 for local agentic coding – VentureBeat
- Open source Xiaomi MiMo-V2.5 and V2.5-Pro are among the most efficient (and affordable) at agentic ‘claw’ tasks – VentureBeat
- DeepSeek-V4 arrives with near state-of-the-art intelligence at 1/6th the cost of Opus 4.7, GPT-5.5 – VentureBeat
- One tool call to rule them all? New open source Python tool Runpod Flash eliminates containers for faster AI dev – VentureBeat






