Find Investable Startups and Competitors
Search thousands of startups using natural language—just describe what you're looking for
Top 50 Ai Compute Platform - Late Stage
Discover the top 50 Ai Compute Platform startups at Late Stage. Browse funding data, key metrics, and company insights. Average funding: $228.5M.
Sort by
Voltage Park
Provides bare-metal access to high-performance AI compute infrastructure powered by NVIDIA HGX H100 GPUs and a 3200 Gbps InfiniBand network, enabling low-latency, scalable training and inference for large-scale machine learning models. Offers transparent pricing and flexible deployment options, including on-demand nodes and long-term contracts, to meet the needs of demanding workloads in AI, HPC, and real-time applications.
Funding: $500M+
Rough estimate of the amount of funding raised
Nscale
Nscale provides a GPU cloud platform optimized for AI workloads, featuring on-demand compute and inference services, dedicated training clusters, and scalable GPU nodes. The platform addresses the high costs and inefficiencies associated with AI model training and deployment by offering a fully integrated infrastructure powered by renewable energy in Europe.
Funding: $100M+
Rough estimate of the amount of funding raised
Nebius AI
Provides a fully managed AI cloud platform powered by NVIDIA® H100 and H200 Tensor Core GPUs, offering scalable GPU clusters with InfiniBand networking for high-speed data processing. Enables efficient model training, fine-tuning, and inference with tools like MLflow, PostgreSQL, and Apache Spark, reducing the complexity and cost of deploying AI applications at scale.
Funding: $500M+
Rough estimate of the amount of funding raised
QScale
The startup develops high-density computing infrastructure specifically designed for AI processing, enabling efficient operationalization of machine learning and compute-intensive workloads. Their platform offers environmentally responsible data processing, allowing clients to achieve faster results while minimizing carbon emissions.
Funding: $100M+
Rough estimate of the amount of funding raised
Crusoe
Crusoe provides a managed AI cloud platform that delivers low‑latency, high‑throughput inference for large‑context models using NVIDIA and AMD GPUs with its MemoryAlloy engine. The service abstracts cluster provisioning via an API‑key workflow, auto‑scales on Kubernetes/Slurm, and includes a web console for one‑click model deployment, while its renewable‑powered data centers reduce compute costs by up to 80 %.
Funding: $1B+
Rough estimate of the amount of funding raised
Anyscale
Anyscale provides a configurable AI platform powered by RayTurbo, enabling developers to optimize and scale AI applications across any cloud and hardware configuration. The platform enhances GPU utilization and reduces cloud costs by up to 50%, facilitating faster model training and deployment for complex AI workloads.
Funding: $200M+
Rough estimate of the amount of funding raised
Foundry
Foundry provides a cloud-native platform offering on-demand access to NVIDIA GPUs across multiple clouds without long-term contracts. The platform integrates with Kubernetes, enabling AI developers to flexibly reserve compute or bid on spot capacity for training and inference workloads. This model allows users to match GPU supply precisely to demand while optimizing costs through usage-based billing and resalable reservations.
Funding: $50M+
Rough estimate of the amount of funding raised
Lambda
Lambda provides an on‑demand supercomputing platform that lets AI teams provision private, single‑tenant GPU clusters with the latest NVIDIA GB300, B200, and H200 accelerators via a web console or API. The service offers up to 64‑GPU nodes with NVLink and InfiniBand interconnects, SOC 2 Type II security, and pay‑as‑you‑go per‑GPU‑hour billing, enabling scalable training and inference for research labs and enterprise ML teams.
Funding: $200M+
Rough estimate of the amount of funding raised
CoreWeave
CoreWeave provides a Kubernetes-native cloud platform optimized for large-scale, GPU-accelerated workloads, offering on-demand access to a diverse range of NVIDIA GPUs and CPU instances. This infrastructure enables businesses to achieve up to 35 times faster performance and 80% cost savings compared to traditional cloud providers, addressing the need for efficient and scalable compute resources in AI and machine learning applications.
Funding: $500M+
Rough estimate of the amount of funding raised
4Paradigm
4Paradigm provides an AI enablement platform that delivers industry‑specific large models built from multi‑modal data and a software‑defined compute layer that abstracts hardware for high‑throughput, low‑cost processing. The platform includes AutoML, transfer‑learning tools, and a generative‑AI development suite that automates model creation, code generation, review, and deployment, all delivered via secure, GDPR‑compliant cloud services.
Funding: $100M+
Rough estimate of the amount of funding raised
Saronic
Saronic offers a cloud‑native AI platform that centralizes the full machine‑learning lifecycle for enterprise teams. It provides auto‑scaling compute for distributed training, automated data‑ingestion and feature‑store pipelines, version‑controlled model management, and secure inference APIs with built‑in explainability and audit logging. The platform integrates with major data warehouses, enabling data‑science and analytics groups to deploy predictive models at scale while maintaining governance and compliance.
Funding: $500M+
Rough estimate of the amount of funding raised
Vultr
Vultr provides cloud infrastructure with dedicated clusters and on-demand virtual machines powered by AMD and NVIDIA GPUs, enabling efficient deployment of AI and high-performance computing workloads. The platform offers scalable solutions at competitive pricing, addressing the need for accessible and powerful computing resources for developers and businesses globally.
Funding: $200M+
Rough estimate of the amount of funding raised
Ori Industries
Ori provides on-demand access to top-tier GPUs and serverless Kubernetes for training and deploying machine learning models at scale. The platform offers cost-optimized solutions that allow users to pay only for the resources they utilize, addressing the need for flexible and efficient AI infrastructure.
Funding: $100M+
Rough estimate of the amount of funding raised
GMI Cloud
GMI Cloud provides instant access to NVIDIA H100 GPUs for training and deploying generative AI applications, utilizing a Kubernetes-based cluster engine for efficient workload orchestration. This platform addresses the need for rapid GPU provisioning and management, enabling developers to focus on building AI models without the complexities of infrastructure setup.
Funding: $100M+
Rough estimate of the amount of funding raised
Krutrim
Krutrim provides an AI computing infrastructure and AI-powered applications tailored for the Indian market, enabling businesses to leverage machine learning and data analytics. This platform addresses the need for accessible and scalable AI solutions, enhancing operational efficiency and decision-making capabilities for local enterprises.
Funding: $50M+
Rough estimate of the amount of funding raised
Axelera AI
10
Relative Traction Score based on online presence metrics compared to companies in the same age group.
Axelera AI develops and sells high-performance, energy-efficient AI inference hardware for edge devices. Their Metis AI Platform integrates a specialized in-memory computing architecture with a comprehensive software stack, enabling efficient deployment of deep learning models for computer vision and natural language processing applications.
Funding: $50M+
Rough estimate of the amount of funding raised
d-Matrix
D-Matrix has developed Corsair, an AI inference platform that achieves 60,000 tokens per second with 1 ms latency for Llama3 8B models, significantly enhancing throughput and energy efficiency in datacenters. This technology addresses the high computational costs and energy consumption associated with large-scale AI inference, enabling organizations to scale their AI capabilities sustainably.
Funding: $100M+
Rough estimate of the amount of funding raised
EigenCloud
EigenCloud provides verifiable infrastructure for AI inference, general compute, and data availability, embedding cryptographic proofs into each operation. Its EigenAI service delivers deterministic, OpenAI‑compatible LLM inference, while EigenCompute generates succinct execution proofs for arbitrary container workloads, and EigenDA offers a high‑throughput (≈100 MB/s) data availability layer for rollups. Operators can stake ETH and EIGEN on EigenLayer to secure these off‑chain services and earn rewards.
Funding: $50M+
Rough estimate of the amount of funding raised
Together AI
Together AI provides a cloud-based platform for training, fine-tuning, and deploying open-source generative AI models using NVIDIA's latest GPUs, enabling users to achieve high performance at a lower cost. The platform addresses the need for scalable AI infrastructure by offering customizable solutions that support the entire generative AI lifecycle, from model development to production deployment.
Funding: $200M+
Rough estimate of the amount of funding raised
Ceremorphic
Develops a proprietary AI supercomputing architecture using a 3nm process and multi-threaded processing (ThreadArch®) to deliver energy-efficient, high-performance training for large AI models. This technology addresses the growing demands of AI applications and drug discovery by reducing compute costs and energy consumption while enabling faster model training and in silico drug design.
Funding: $50M+
Rough estimate of the amount of funding raised
Taalas
Taalas provides a platform that transforms any AI model into custom silicon, creating Hardcore Models that are hardwired for optimal performance. This approach significantly enhances computational efficiency, achieving up to 1000 times the performance of traditional software implementations.
Funding: $50M+
Rough estimate of the amount of funding raised
Xoda
Xoda AI offers a decentralized platform for AI development and deployment, integrating blockchain, IPFS, and LLMs. It provides an ecosystem for developers to build, train, and monetize AI models and applications, while resource providers can contribute compute power and earn rewards.
Funding: $50M+
Rough estimate of the amount of funding raised
Together AI
Together AI provides a cloud platform that offers serverless OpenAI‑compatible inference APIs for over 200 open‑source models, accelerated up to 4× by its ATLAS runtime. Users can provision on‑demand or reserved NVIDIA GPU clusters for fine‑tuning and batch inference, with per‑token or hourly usage pricing and enterprise‑grade security.
Habana
Habana Labs develops Intel® Gaudi® AI accelerators designed for high-performance deep learning training and inference, providing enterprises and cloud providers with efficient compute solutions. Their technology delivers up to 40% better price/performance on cloud instances, addressing the need for cost-effective and scalable AI infrastructure.
Funding: $50M+
Rough estimate of the amount of funding raised
Enflame
Enflame develops cloud-based deep learning chips specifically designed for AI training platforms, enhancing computational efficiency and speed. This technology addresses the high resource demands of AI model training, enabling faster iterations and reduced operational costs for businesses.
Funding: $200M+
Rough estimate of the amount of funding raised
Etched.ai
Etched.ai develops Sohu, the world's first ASIC specifically designed for transformer models, enabling AI computations to be executed at least ten times faster and more cost-effectively than traditional GPUs. This technology allows for real-time processing of large-scale AI models, enhancing applications such as voice agents and content generation.
Funding: $100M+
Rough estimate of the amount of funding raised
MatX
MatX manufactures specialized hardware designed for training and inference of large AI models, delivering up to 10× more computing power for workloads with over 7 billion parameters. This enables researchers and startups to efficiently train advanced models, significantly reducing the time and cost associated with developing state-of-the-art AI systems.
Funding: $100M+
Rough estimate of the amount of funding raised
Clarifai
Clarifai offers an end-to-end AI lifecycle platform that automates data labeling, model training, and deployment, enabling organizations to build and operationalize AI applications efficiently. By standardizing workflows and optimizing compute resources, the platform reduces development time and costs, allowing enterprises to scale AI solutions rapidly.
Funding: $50M+
Rough estimate of the amount of funding raised
Gensyn
Gensyn is a machine learning compute protocol that connects distributed resources to facilitate the training of deep learning models. This approach addresses the need for open, permissionless, and neutral frameworks that enable efficient scaling and collaboration in machine intelligence development.
Modular
Modular provides a unified AI software development platform that enables developers to optimize and deploy AI models across various cloud environments without rewriting code. By leveraging high-performance Mojo programming and a secure inference stack, the platform significantly reduces cloud costs while enhancing computational efficiency for enterprise applications.
Fireworks AI
Fireworks AI provides a serverless inference platform that enables the rapid deployment and fine-tuning of compound AI models, optimizing for speed and cost efficiency. The technology addresses the challenges of slow model inference and high operational costs, allowing businesses to scale AI applications effectively while maintaining low latency and high throughput.
Esperanto Technologies
Esperanto Technologies develops massively parallel, energy-efficient chips based on the RISC-V instruction set architecture, specifically designed for Generative AI and high-performance computing (HPC) applications. Their ET-SoC-1 chip features over a thousand low-power RISC-V cores, providing superior compute efficiency and significantly reducing total cost of ownership for AI inference and HPC workloads.
Funding: $50M+
Rough estimate of the amount of funding raised
Core Scientific
Core Scientific provides high‑density colocation data centers with a minimum of 30 MW per site and up to 200 kW per cabinet, featuring GPU‑optimized racks, direct liquid‑cooling, and carrier‑neutral high‑bandwidth fiber. The facilities deliver over 1.3 GW of contracted power across the United States, with 24 × 7 NOC monitoring, flexible turnkey provisioning, and managed power services to support AI/ML, hyperscale, and other compute‑intensive workloads. This enables customers to scale high‑performance compute without building their own infrastructure, lowering total cost of ownership and deployment time.
Funding: $500M+
Rough estimate of the amount of funding raised
NextSilicon
NextSilicon's Maverick-2 Intelligent Compute Accelerator (ICA) utilizes software-defined hardware to dynamically optimize performance for high-performance computing (HPC) and artificial intelligence (AI) workloads. This technology eliminates the need for extensive code rewrites, significantly reducing development time and enabling faster insights across various applications.
Funding: $200M+
Rough estimate of the amount of funding raised
Luminous Computing
Luminous Computing develops photonics chips designed to provide the necessary compute, memory, and bandwidth for advanced artificial intelligence applications. This technology addresses the limitations of current hardware, enabling instant processing of complex queries and facilitating the development of next-generation AI solutions.
Funding: $100M+
Rough estimate of the amount of funding raised
NeuReality
NeuReality designs AI-centric infrastructure that integrates a network addressable processing unit (NAPU) with purpose-built software to streamline AI inference workflows. This solution reduces reliance on traditional CPUs and networking components, addressing the complexity and inefficiencies that hinder AI model deployment and scalability.
SiMa.ai
SiMa.ai develops a software-centric platform utilizing its proprietary Machine Learning System on Chip (MLSoC) technology to enable efficient deployment of multimodal AI applications at the edge. This platform addresses the need for high-performance, power-efficient solutions that can scale across various edge devices and applications, significantly improving processing speed and energy consumption.
Funding: $200M+
Rough estimate of the amount of funding raised
Axelera AI
Axelera AI manufactures AI acceleration hardware, specifically the Metis AI Processing Unit (AIPU), designed for efficient edge computing with up to 214 TOPS performance and 15 TOPS per watt. The technology addresses the need for cost-effective and energy-efficient solutions in generative AI and computer vision applications across various industries, including retail and security.
Funding: $100M+
Rough estimate of the amount of funding raised
Panmnesia
The startup manufactures a chip that utilizes Compute Express Link technology to enable data center operators to efficiently pool and manage artificial intelligence accelerators, processors, and memory. This approach enhances system performance by providing adequate memory resources for diverse device integration, addressing the challenges of scalability and resource allocation in large-scale computing environments.
Funding: $50M+
Rough estimate of the amount of funding raised
Petuum
The startup offers a machine learning infrastructure platform that provides a flexible operating system and virtualization interface for building and deploying machine learning and deep learning applications at scale. This technology enables enterprises to manage applications and hardware from a single terminal, resulting in increased productivity, reduced operational costs, and faster delivery times.
Armada
Armada provides an edge computing platform that integrates connectivity, ruggedized mobile data centers, and real-world AI applications to enable real-time data processing in remote environments. This technology addresses challenges in industries such as oil and gas, manufacturing, and logistics by enhancing safety, automating operations, and improving decision-making capabilities.
Funding: $50M+
Rough estimate of the amount of funding raised
DEEPX
Develops on-device AI semiconductor solutions, including custom NPUs, SoC ASICs, and specialized modules, optimized for low power consumption and high performance in applications like video analytics, security, and robotics. By enabling real-time AI processing with support for multiple models on a single chip, DEEPX addresses the challenges of latency, privacy, and network costs associated with cloud-based systems. Its scalable architecture and 259 patents ensure cost-competitive, silicon-proven products for global markets.
Funding: $100M+
Rough estimate of the amount of funding raised
Baidu
Baidu provides an integrated AI ecosystem comprising a cloud‑based AI Open Platform with over 270 pre‑trained model APIs for vision, speech, and language, the DuerOS voice‑assistant SDK for multimodal interaction, and the Apollo autonomous‑driving stack offering perception, planning, and safety‑critical tools. These services run on Baidu’s Kunlun AI chips and the PaddlePaddle deep‑learning framework, delivering scalable, production‑grade performance and pay‑as‑you‑go pricing for developers, enterprise IT teams, and automotive OEMs.
Funding: $500M+
Rough estimate of the amount of funding raised
webAI
webAI provides an enterprise platform for on-device AI development and deployment, enabling organizations to build, train, and run custom models locally. This approach ensures full data sovereignty, reduces latency, and lowers operational costs by eliminating cloud dependencies.
Funding: $50M+
Rough estimate of the amount of funding raised
Cornelis Networks
Cornelis provides high-performance fabrics specifically designed for AI infrastructure, ensuring universal compatibility with accelerators and GPUs while delivering high bandwidth and scalable architecture. This technology meets the critical demands of commercial, scientific, academic, and government organizations operating in hyperscale, cloud AI, and on-premises AIHPC environments.
Funding: $100M+
Rough estimate of the amount of funding raised
Baseten
Baseten provides a platform for deploying and serving machine learning models with optimized inference speed and autoscaling capabilities, enabling seamless transition from development to production. The solution addresses the complexities of model infrastructure management, allowing teams to focus on building and iterating on their AI applications without incurring excessive costs.
Funding: $50M+
Rough estimate of the amount of funding raised
Lightelligence
Develops photonic computing solutions that integrate optical and electronic components to accelerate AI workloads, addressing the limitations of traditional electronic systems, such as the "memory wall." Products like the HUMMINGBIRD optical network-on-chip and PACE photonic arithmetic engine enable exponential increases in processing speed and efficiency for domain-specific applications.
Funding: $200M+
Rough estimate of the amount of funding raised
Groq
Groq accelerates AI inference with custom-designed Language Processing Units (LPUs) that deliver sub-millisecond latency and consistent performance. Their cloud platform and on-premise solutions enable developers to deploy AI models efficiently and cost-effectively.
DreamBig Semiconductor
DreamBig Semiconductor offers a Chiplet platform with SMARTNIC-DPU solutions designed for low latency and high throughput in AI, data centers, and storage acceleration. Their technology addresses the need for efficient data processing and inherent security in high-demand computing environments.
Funding: $50M+
Rough estimate of the amount of funding raised
Hadean
The startup offers a cloud-distributed computing platform that enables developers to build, run, and scale real-time applications through a command-line interface. Its infrastructure allows for the deployment of algorithms at scale, facilitating the creation and monetization of Metaverse experiences for businesses.
Funding: $50M+
Rough estimate of the amount of funding raised