Key Highlights
- CoreWeave stock increased approximately 6% during premarket hours following the announcement of a multi-year infrastructure partnership with Perplexity AI.
- Perplexity chose CoreWeave to serve as a backend cloud infrastructure provider, utilizing NVIDIA GB200 NVL72 GPU clusters for inference operations.
- The cloud infrastructure will power Perplexity’s Sonar platform and Search API offerings as the AI search company scales its operations.
- Perplexity has initiated workload deployment using CoreWeave’s Kubernetes-based services within its broader multi-cloud infrastructure approach.
- CoreWeave plans to implement Perplexity Enterprise Max across its organization, giving staff access to AI-enhanced research and search capabilities.
CoreWeave unveiled a multi-year partnership to deliver inference computing infrastructure for Perplexity AI, a rising AI search platform supported by investors including Jeff Bezos and Nvidia. The announcement drove CoreWeave shares upward during Wednesday’s premarket session. Both organizations confirmed that CoreWeave will supply dedicated GPU cloud resources to support Perplexity’s growing portfolio of search and API offerings.
Perplexity Partnership Drives CoreWeave Stock Higher in Premarket Session
CoreWeave experienced a share price increase of nearly six percent during premarket activity after revealing the infrastructure collaboration with Perplexity. This arrangement establishes CoreWeave as a core cloud infrastructure provider for the AI-powered search company.
Both parties disclosed that Perplexity will execute inference computing tasks on NVIDIA GB200 NVL72 cluster systems operated by CoreWeave. These advanced clusters deliver the computational power required for cutting-edge AI applications.
Perplexity plans to operate production-level inference workloads on exclusive GPU resources distributed throughout CoreWeave’s cloud infrastructure. This system will accommodate increasing usage from Perplexity’s Sonar platform and Search API solutions.
Max Hjelm, senior vice president of revenue at CoreWeave, outlined the essential components for contemporary AI implementations. He stated, “AI applications running in production require more than raw infrastructure and demand performance, reliability, and a cloud platform designed for AI.”
Hjelm emphasized that CoreWeave focuses on streamlining computational processes for AI developers creating production-ready systems. The platform was engineered specifically to accommodate extensive machine learning operations.
Perplexity emphasized that inference speed remains critical to delivering quality user experiences in AI search applications. Rapid response generation and consistent latency directly impact result delivery speed.
The partnership establishes CoreWeave technology as a cornerstone for upcoming service enhancements throughout Perplexity’s product suite. Both companies verified that inference operations have commenced on CoreWeave’s infrastructure.
Cloud Infrastructure Partnership Enables Perplexity Platform Scaling
Perplexity leadership highlighted CoreWeave’s technical proficiency when discussing the partnership selection. Dmitry Shevelenko, chief business officer at Perplexity, commended the provider’s engineering capabilities and support structure.
Shevelenko expressed appreciation for CoreWeave’s specialized knowledge and collaborative methodology with AI enterprises. He indicated the collaboration would drive improvements in infrastructure optimization and accelerate product innovation.
Perplexity verified that workloads currently operate through CoreWeave’s Kubernetes platform for containerized application management. The company additionally utilizes W&B Models for model training and refinement as part of its comprehensive multi-cloud architecture.
This implementation facilitates the expansion of AI search infrastructure processing substantial query volumes. Inference operations apply trained models to produce responses, suggestions, and search outcomes.
CoreWeave emphasized that specialized GPU cloud infrastructure has become essential for AI organizations managing expanding computational demands. The provider demonstrated impressive performance in MLPerf benchmark testing for machine learning capabilities.
The organization also achieved platinum status in SemiAnalysis ClusterMAX evaluations measuring performance and reliability. These assessments evaluate computational effectiveness and operational consistency across extensive GPU cluster deployments.



