Blaize Secures $50M NeoTensr Deal While Reaffirming $130M Full-Year Outlook
Q1 revenue expected at $2.7M amid supply constraints, with strong demand and new AI infrastructure partnership driving growth into 2026
Blaize delivers production-ready, application-level AI services that enable cloud providers, system integrators, and enterprises to deploy business-outcome applications without building the underlying AI stack from scratch. A unified platform built on programmable silicon and a composable software stack packages multimodal inference, business logic, management, and orchestration into modular APIs — delivering the majority of required functionality out of the box.
The result is a faster path from pilot to production at lower cost per query, with Forward Deployed Engineering closing the last mile to revenue-generating AI operations.
vs. GPU-Only
vs. GPU-Only
for CSP Partners
The application layer is where enterprise AI stalls. Hundreds of specialized vendors deliver narrow inference functions, each adding its own procurement, integration, and operational overhead.
That fragmentation compounds a cost structure already working against scale: inference consumes 80% of enterprise AI budgets, and cheaper unit costs haven’t lowered total spending — they’ve raised it.
The market is done assembling parts. It’s converging on hybrid AI platforms.
The Structural Shift
>
>
>
>
>
One platform architected for heterogeneity — mixed computing platforms, mixed architectures, brownfield environments — packaging multimodal inference, business logic, orchestration, and lifecycle management into modular APIs.
Application-Level AI Services via API
Vision | Video | Document | Multimodal | Speech | Moderation
Application Use Cases
Smarty City Analytics | Retail Intelligence | Industrial Monitoring | Security & Surveillance
Intelligent Software
Integration & Runtime Enablement | AI Services Engine | API Gateway
Programmable Silicon & Hybrid AI Compute
Blaize GSP® + GPU | Cards | Accelerators v Complete Systems
Application-level APIs package inference, business logic, provisioning, scheduling, and lifecycle management into composable services. Launch differentiated AI offerings without building the platform from scratch — the majority of required functionality is delivered out of the box.
Programmable silicon delivers multimodal AI at lower cost per query than GPU-only approaches. 2.4× more video streams per rack and ~60% less power per rack means better margins as you scale — with intelligent workload routing that monetizes existing GPU capacity.
Deploy production-ready AI as application-level APIs without assembling fragmented point solutions or standing up internal AI engineering teams. Modular API architecture delivers the majority of required functionality out of the box — with sovereign deployment, data residency, and open integration built in.
Blaize operates alongside GPUs and third-party infrastructure under a unified software layer. Standard APIs and developer tools connect to existing sensors, VMS systems, and enterprise applications. No proprietary interfaces. No platform replacement required.
Deployment-ready systems for commercial security, smart retail, smart traffic, defense, industrial, and energy — bundling hardware, software, SDK, and vertical applications into validated, deployable packages without the R&D costs.
Run AI on your own infrastructure with full control over data residency, compliance, and operational sovereignty. Each instance generates operational telemetry that feeds model optimization and workload routing — a platform that delivers more value at site 100 than it did at site 1.
See how Blaize closes the gap between AI pilot and revenue-generating instance — in under three minutes.

South Asia
Smart public safety AI across distributed urban infrastructure environment with full operational autonomy.

Middle East/The Gulf Region
Full data residency and sovereign AI capability without hyperscaler dependency.

Asia Pacific
Multi-environments spanning smart cities, industrial automation, and public services from edge to cloud.
Q1 revenue expected at $2.7M amid supply constraints, with strong demand and new AI infrastructure partnership driving growth into 2026
Helps providers and enterprises deploy application-level AI services faster, lower cost per query, and unlock recurring AI revenue
Expanded collaboration introduces validated reference architecture and Innovation Hub to drive real-world AI instances across edge and data centers in Asia-Pacific.
Hybrid AI inference uses a combination of specialized silicon (like the Blaize Graph Streaming Processor) and GPUs to run AI models in production. An intelligent workload scheduler routes each query to the optimal processor based on cost, power, and performance targets. This delivers significantly lower cost per inference and power consumption compared to GPU-only approaches — 2.4× more video streams per rack and 60% less power per rack — making always-on AI economically viable at distributed scale.
Blaize delivers application-level AI services through modular, composable APIs — with the majority of required functionality delivered out of the box. Cloud providers launch differentiated AI offerings without building the stack from scratch. Programmable silicon delivers more inference per watt, Forward Deployed Engineering handles the last mile to production through a revenue-share model, and the composable software stack compresses time to AI service revenue.
Blaize serves two primary audiences: infrastructure providers (cloud service providers, data center operators, colocation facilities, telcos, system integrators) and enterprise consumers (government agencies, defense organizations, smart city programs, retail, industrial, and energy operations). Deployments span from far-edge environments to private data centers to hybrid cloud.
The Blaize GSP is a purpose-built AI inference processor with a graph-native architecture. It executes multiple models concurrently without batching, using task-level parallelism to eliminate redundant memory transfers. This dramatically reduces power and compute overhead per inference. The GSP operates across form factors — from data center inference cards to ruggedized edge accelerators — with consistent performance characteristics.
Inference platforms deliver optimized model execution — the enterprise assembles the application. Cloud development platforms provide the infrastructure to build — the enterprise architects and manages the build. Blaize delivers application-level AI services: production-ready APIs that package inference, business logic, management, and orchestration into modular services, with Forward Deployed Engineering closing the last mile to production. Production-ready enough that the buyer doesn’t build from scratch, but composable enough that providers customize, extend, and brand the services for their own markets.
Forward Deployed Engineering places Blaize engineers directly into customer and partner environments to handle application integration, workflow configuration, ongoing optimization, and the last-mile customization that takes application-level AI services from pre-built platform capability to production-scale deployment. For CSP and infrastructure partners, it’s delivered through a revenue-share model — Blaize’s commercial success is tied directly to the AI service revenue generated on the platform. A partnership with skin in the game, not a consulting engagement.
Yes. Blaize is open and hybrid by design. The platform operates alongside GPUs and third-party infrastructure under a unified software layer. Standard APIs and a composable software stack enable integration across existing infrastructure, sensors, VMS systems, and third-party platforms without proprietary lock-in or rip-and-replace requirements.
Application-level AI services package multimodal inference, business logic, scheduling, lifecycle management, and orchestration into modular APIs that deliver production-ready AI functionality — not isolated inference functions. Instead of returning a raw model output like a bounding box or classification, application-level APIs encapsulate the full stack required to run an AI application in production: identity management, provisioning, compliance controls, and operational telemetry. The result is a faster path from pilot to production at lower cost per query.
Blaize delivers the platform,
the APIs, and the Forward Deployed Engineering to get you to production.