Find Investable Startups and Competitors
Search thousands of startups using natural language—just describe what you're looking for
Top 50 Ai Compute Platform - Series B
Discover the top 50 Ai Compute Platform startups at Series B. Browse funding data, key metrics, and company insights. Average funding: $71.3M.
Sort by
FlexAI
FlexAI provides a universal AI compute platform that enables developers to run AI workloads across diverse hardware architectures without code modifications. This approach maximizes resource utilization and energy efficiency, reducing operational complexity and minimizing failures in AI product development.
Funding: $20M+
Rough estimate of the amount of funding raised
TensorWave
TensorWave provides a cloud platform optimized for AI workloads, utilizing AMD's Instinct MI300X accelerators for enhanced training, fine-tuning, and inference capabilities. The platform offers immediate availability, lower total cost of ownership, and seamless integration with popular frameworks like PyTorch and TensorFlow, addressing the need for efficient and scalable AI compute solutions.
Funding: $20M+
Rough estimate of the amount of funding raised
Nscale
Nscale provides a GPU cloud platform optimized for AI workloads, featuring on-demand compute and inference services, dedicated training clusters, and scalable GPU nodes. The platform addresses the high costs and inefficiencies associated with AI model training and deployment by offering a fully integrated infrastructure powered by renewable energy in Europe.
Funding: $100M+
Rough estimate of the amount of funding raised
QScale
The startup develops high-density computing infrastructure specifically designed for AI processing, enabling efficient operationalization of machine learning and compute-intensive workloads. Their platform offers environmentally responsible data processing, allowing clients to achieve faster results while minimizing carbon emissions.
Funding: $100M+
Rough estimate of the amount of funding raised
Foundry
Foundry provides a cloud-native platform offering on-demand access to NVIDIA GPUs across multiple clouds without long-term contracts. The platform integrates with Kubernetes, enabling AI developers to flexibly reserve compute or bid on spot capacity for training and inference workloads. This model allows users to match GPU supply precisely to demand while optimizing costs through usage-based billing and resalable reservations.
Funding: $50M+
Rough estimate of the amount of funding raised
EnCharge AI
EnCharge AI develops a scalable analog in-memory computing platform that enhances AI performance by achieving 20 times higher efficiency and 10 times lower total cost of ownership compared to traditional GPU solutions. This technology enables on-device processing, significantly reducing CO2 emissions and ensuring data privacy while making advanced AI accessible beyond cloud infrastructure.
Funding: $20M+
Rough estimate of the amount of funding raised
Exabits.ai
The startup provides a decentralized infrastructure for AI that enables users to access and purchase computing services, data storage, and Model as a Service (MaaS) at scale. This platform addresses the high costs and inefficiencies associated with deploying sophisticated AI solutions, allowing users to optimize performance and resource utilization.
Funding: $20M+
Rough estimate of the amount of funding raised
RunPod
RunPod is a cloud platform that provides globally distributed GPU resources for deploying and scaling machine learning applications, enabling developers to run AI workloads without managing infrastructure. The platform reduces cold-start times to under 250 milliseconds and offers flexible pricing, allowing users to efficiently handle fluctuating demand while minimizing operational costs.
Funding: $20M+
Rough estimate of the amount of funding raised
Genesis Cloud
Genesis Cloud provides a GPU cloud platform built on NVIDIA's reference architecture, delivering up to 35 times more performance for AI and machine learning workloads at 80% lower costs compared to traditional cloud providers. The platform ensures high security and compliance with EU regulations, enabling enterprises to efficiently manage and scale their AI applications.
Funding: $20M+
Rough estimate of the amount of funding raised
4Paradigm
4Paradigm provides an AI enablement platform that delivers industry‑specific large models built from multi‑modal data and a software‑defined compute layer that abstracts hardware for high‑throughput, low‑cost processing. The platform includes AutoML, transfer‑learning tools, and a generative‑AI development suite that automates model creation, code generation, review, and deployment, all delivered via secure, GDPR‑compliant cloud services.
Funding: $100M+
Rough estimate of the amount of funding raised
Ori Industries
Ori provides on-demand access to top-tier GPUs and serverless Kubernetes for training and deploying machine learning models at scale. The platform offers cost-optimized solutions that allow users to pay only for the resources they utilize, addressing the need for flexible and efficient AI infrastructure.
Funding: $100M+
Rough estimate of the amount of funding raised
GMI Cloud
GMI Cloud provides instant access to NVIDIA H100 GPUs for training and deploying generative AI applications, utilizing a Kubernetes-based cluster engine for efficient workload orchestration. This platform addresses the need for rapid GPU provisioning and management, enabling developers to focus on building AI models without the complexities of infrastructure setup.
Funding: $100M+
Rough estimate of the amount of funding raised
Krutrim
Krutrim provides an AI computing infrastructure and AI-powered applications tailored for the Indian market, enabling businesses to leverage machine learning and data analytics. This platform addresses the need for accessible and scalable AI solutions, enhancing operational efficiency and decision-making capabilities for local enterprises.
Funding: $50M+
Rough estimate of the amount of funding raised
io.net
Io.net Cloud offers a decentralized computing network that provides machine learning engineers with instant, permissionless access to global GPU resources for their workloads. This platform enables efficient deployment of pre-configured clusters, significantly reducing costs and deployment time for AI startups.
Funding: $20M+
Rough estimate of the amount of funding raised
Axelera AI
10
Relative Traction Score based on online presence metrics compared to companies in the same age group.
Axelera AI develops and sells high-performance, energy-efficient AI inference hardware for edge devices. Their Metis AI Platform integrates a specialized in-memory computing architecture with a comprehensive software stack, enabling efficient deployment of deep learning models for computer vision and natural language processing applications.
Funding: $50M+
Rough estimate of the amount of funding raised
d-Matrix
D-Matrix has developed Corsair, an AI inference platform that achieves 60,000 tokens per second with 1 ms latency for Llama3 8B models, significantly enhancing throughput and energy efficiency in datacenters. This technology addresses the high computational costs and energy consumption associated with large-scale AI inference, enabling organizations to scale their AI capabilities sustainably.
Funding: $100M+
Rough estimate of the amount of funding raised
Mirantis
Mirantis' k0rdent platform automates provisioning of GPU‑optimized, multi‑tenant environments on bare‑metal and Kubernetes, delivering on‑demand AI infrastructure in seconds. It includes a self‑service marketplace, DPU‑based isolation, and a GitOps‑driven control plane for managing hybrid and multi‑cloud clusters, enabling enterprises and service providers to scale AI workloads and monetize idle GPU capacity.
Funding: $20M+
Rough estimate of the amount of funding raised
EigenCloud
EigenCloud provides verifiable infrastructure for AI inference, general compute, and data availability, embedding cryptographic proofs into each operation. Its EigenAI service delivers deterministic, OpenAI‑compatible LLM inference, while EigenCompute generates succinct execution proofs for arbitrary container workloads, and EigenDA offers a high‑throughput (≈100 MB/s) data availability layer for rollups. Operators can stake ETH and EIGEN on EigenLayer to secure these off‑chain services and earn rewards.
Funding: $50M+
Rough estimate of the amount of funding raised
Sahara AI
Sahara AI provides an AI‑native blockchain platform that combines curated data services, on‑demand decentralized compute, and a marketplace for AI assets. It records immutable on‑chain provenance for datasets, models, and agents, uses the $SAHARA token for licensing, per‑inference payments and automatic royalty distribution, and offers SOC2‑certified security. The solution enables model developers, enterprise AI teams, and research labs to access trusted data, scalable compute, and a secure monetization layer while reducing intermediaries.
Funding: $20M+
Rough estimate of the amount of funding raised
Aethir
Aethirs provides a decentralized cloud infrastructure that delivers on-demand access to enterprise-grade GPUs for AI model training and real-time gaming applications. This solution addresses the need for scalable, low-latency compute resources while ensuring high performance and security across a global network.
Funding: $20M+
Rough estimate of the amount of funding raised
Deep Infra
Provides a serverless machine learning inference platform that enables businesses to deploy and scale AI models via a simple API, eliminating the need for complex ML infrastructure. It reduces costs and improves efficiency by offering pay-per-use pricing, low-latency performance, and automatic scaling on dedicated A100 and H100 GPUs.
Funding: $20M+
Rough estimate of the amount of funding raised
Edgecortix
EdgeCortix develops the SAKURA-II Edge AI Platform, an energy-efficient AI accelerator that delivers up to 240 TOPS for real-time inferencing in compact, low-power modules. This technology addresses the need for high-performance AI processing at the edge, significantly reducing operational costs across various sectors, including defense, robotics, and smart manufacturing.
Funding: $20M+
Rough estimate of the amount of funding raised
Ceremorphic
Develops a proprietary AI supercomputing architecture using a 3nm process and multi-threaded processing (ThreadArch®) to deliver energy-efficient, high-performance training for large AI models. This technology addresses the growing demands of AI applications and drug discovery by reducing compute costs and energy consumption while enabling faster model training and in silico drug design.
Funding: $50M+
Rough estimate of the amount of funding raised
Taalas
Taalas provides a platform that transforms any AI model into custom silicon, creating Hardcore Models that are hardwired for optimal performance. This approach significantly enhances computational efficiency, achieving up to 1000 times the performance of traditional software implementations.
Funding: $50M+
Rough estimate of the amount of funding raised
Xoda
Xoda AI offers a decentralized platform for AI development and deployment, integrating blockchain, IPFS, and LLMs. It provides an ecosystem for developers to build, train, and monetize AI models and applications, while resource providers can contribute compute power and earn rewards.
Funding: $50M+
Rough estimate of the amount of funding raised
Habana
Habana Labs develops Intel® Gaudi® AI accelerators designed for high-performance deep learning training and inference, providing enterprises and cloud providers with efficient compute solutions. Their technology delivers up to 40% better price/performance on cloud instances, addressing the need for cost-effective and scalable AI infrastructure.
Funding: $50M+
Rough estimate of the amount of funding raised
Etched.ai
Etched.ai develops Sohu, the world's first ASIC specifically designed for transformer models, enabling AI computations to be executed at least ten times faster and more cost-effectively than traditional GPUs. This technology allows for real-time processing of large-scale AI models, enhancing applications such as voice agents and content generation.
Funding: $100M+
Rough estimate of the amount of funding raised
MatX
MatX manufactures specialized hardware designed for training and inference of large AI models, delivering up to 10× more computing power for workloads with over 7 billion parameters. This enables researchers and startups to efficiently train advanced models, significantly reducing the time and cost associated with developing state-of-the-art AI systems.
Funding: $100M+
Rough estimate of the amount of funding raised
Hivenet
Hive is a distributed cloud storage and computing platform that utilizes unused digital space and computing power from devices worldwide to provide secure and sustainable data storage and processing capabilities. By reducing reliance on traditional data centers, Hive offers a cost-effective solution that lowers carbon emissions by 77% while enabling users to rent computing resources for AI and high-performance computing tasks.
Funding: $20M+
Rough estimate of the amount of funding raised
Positron
Provides a transformer inference server that delivers up to 5.2x higher performance and 75% lower cost per token compared to Nvidia DGX-H100 systems, optimizing AI model deployment for power-constrained environments. The platform supports seamless integration with HuggingFace models and offers a managed inference service for remote evaluation, enabling efficient scaling and reduced operational expenses for AI-driven applications.
Funding: $20M+
Rough estimate of the amount of funding raised
Clarifai
Clarifai offers an end-to-end AI lifecycle platform that automates data labeling, model training, and deployment, enabling organizations to build and operationalize AI applications efficiently. By standardizing workflows and optimizing compute resources, the platform reduces development time and costs, allowing enterprises to scale AI solutions rapidly.
Funding: $50M+
Rough estimate of the amount of funding raised
Qubrid AI
The startup provides a hybrid cloud platform that combines GPU cloud services with AI controller software for on-premises deployments, enabling enterprises to efficiently manage AI workloads. This solution allows businesses to optimize their data center resources while seamlessly scaling to the public cloud as needed.
Funding: $20M+
Rough estimate of the amount of funding raised
Gensyn
Gensyn is a machine learning compute protocol that connects distributed resources to facilitate the training of deep learning models. This approach addresses the need for open, permissionless, and neutral frameworks that enable efficient scaling and collaboration in machine intelligence development.
Modular
Modular provides a unified AI software development platform that enables developers to optimize and deploy AI models across various cloud environments without rewriting code. By leveraging high-performance Mojo programming and a secure inference stack, the platform significantly reduces cloud costs while enhancing computational efficiency for enterprise applications.
Fireworks AI
Fireworks AI provides a serverless inference platform that enables the rapid deployment and fine-tuning of compound AI models, optimizing for speed and cost efficiency. The technology addresses the challenges of slow model inference and high operational costs, allowing businesses to scale AI applications effectively while maintaining low latency and high throughput.
Esperanto Technologies
Esperanto Technologies develops massively parallel, energy-efficient chips based on the RISC-V instruction set architecture, specifically designed for Generative AI and high-performance computing (HPC) applications. Their ET-SoC-1 chip features over a thousand low-power RISC-V cores, providing superior compute efficiency and significantly reducing total cost of ownership for AI inference and HPC workloads.
Funding: $50M+
Rough estimate of the amount of funding raised
Neural Magic
Neural Magic provides an enterprise inference server solution that optimizes the deployment of open-source large language models (LLMs) on both CPU and GPU infrastructures. By enhancing computational efficiency and reducing hardware requirements, the platform enables organizations to run AI models securely and cost-effectively across various environments, including cloud and edge.
Funding: $20M+
Rough estimate of the amount of funding raised
Wallaroo.AI
The startup offers a cloud-based data processing AI platform that enables the deployment of real-time applications without infrastructure constraints. Its software allows data engineers and architects to efficiently process large data volumes, enhancing outpatient monitoring and real-time bidding while minimizing investment costs.
Funding: $20M+
Rough estimate of the amount of funding raised
Luminous Computing
Luminous Computing develops photonics chips designed to provide the necessary compute, memory, and bandwidth for advanced artificial intelligence applications. This technology addresses the limitations of current hardware, enabling instant processing of complex queries and facilitating the development of next-generation AI solutions.
Funding: $100M+
Rough estimate of the amount of funding raised
Lumen Orbit (YC S24
<problem>
Traditional terrestrial data centers face limitations in scalability, energy costs, and deployment speed due to permitting constraints and reliance on terrestrial power grids. The increasing demand for AI training and high-performance computing requires data centers to scale to gigawatt levels, which is often infeasible on Earth due to energy and infrastructure limitations.
</problem>
<solution>
Lumen Orbit is developing megawatt-scale data centers in space that leverage 24/7 solar energy and passive cooling to overcome the limitations of terrestrial data centers. By deploying data centers in a low-Earth, sun-synchronous orbit, Lumen Orbit can achieve gigawatt scalability without terrestrial permitting constraints. The space-based data centers utilize modular designs for rapid deployment and can be scaled linearly to meet the growing demands of AI training and data processing. This approach reduces operating expenses by using low-cost solar energy and efficient radiative cooling, while also minimizing environmental impact by eliminating freshwater usage and reducing greenhouse gas emissions.
</solution>
<features>
- 24/7 solar power generation with capacity factors exceeding 95%
- Passive radiative cooling using deployable radiators with high emissivity
- Modular design for scalable compute, power, and cooling infrastructure
- High-voltage DC power distribution for efficient energy transfer
- Radiation shielding to protect sensitive components
- Optical (laser) communications for high-throughput and secure data transfer
- Compatibility with heavy-lift launch vehicles for cost-effective deployment
- Use of thin-film silicon solar cells for mass and volume efficiency
</features>
<target_audience>
The primary target audience includes organizations involved in AI training, high-performance computing, and data processing that require scalable and cost-effective data center solutions.
</target_audience>
NeuReality
NeuReality designs AI-centric infrastructure that integrates a network addressable processing unit (NAPU) with purpose-built software to streamline AI inference workflows. This solution reduces reliance on traditional CPUs and networking components, addressing the complexity and inefficiencies that hinder AI model deployment and scalability.
Oort
Oort is a decentralized cloud computing platform that utilizes a blockchain-based verification layer to integrate global resources for secure data storage and AI model training. By leveraging idle computing power from data centers and edge devices, Oort reduces costs by up to 80% while ensuring 100% privacy for all data services.
Funding: $20M+
Rough estimate of the amount of funding raised
Axelera AI
Axelera AI manufactures AI acceleration hardware, specifically the Metis AI Processing Unit (AIPU), designed for efficient edge computing with up to 214 TOPS performance and 15 TOPS per watt. The technology addresses the need for cost-effective and energy-efficient solutions in generative AI and computer vision applications across various industries, including retail and security.
Funding: $100M+
Rough estimate of the amount of funding raised
SynthBee
SynthBee is developing a computing intelligence platform that utilizes advanced algorithms to enhance data processing and decision-making capabilities. This technology enables organizations to increase innovation efficiency while ensuring data security and scalability in their operations.
Funding: $20M+
Rough estimate of the amount of funding raised
Panmnesia
The startup manufactures a chip that utilizes Compute Express Link technology to enable data center operators to efficiently pool and manage artificial intelligence accelerators, processors, and memory. This approach enhances system performance by providing adequate memory resources for diverse device integration, addressing the challenges of scalability and resource allocation in large-scale computing environments.
Funding: $50M+
Rough estimate of the amount of funding raised
XGEN AI
The startup offers a composable AI cloud platform that enables eCommerce teams to create, configure, and deploy tailored AI solutions without requiring prior AI expertise. This platform addresses the challenge of outdated technology in eCommerce by providing a user-friendly toolkit that allows brands to rapidly implement AI to enhance conversion rates and optimize margins.
Funding: $20M+
Rough estimate of the amount of funding raised
Expedera
Provides scalable neural processor unit (NPU) semiconductor IP with a packet-based architecture that enables parallel execution of AI workloads, achieving up to 90% processor utilization. This approach reduces memory overhead, power consumption, and latency while supporting complex AI models across edge devices in industries like mobile, automotive, and industrial automation.
Funding: $20M+
Rough estimate of the amount of funding raised
Petuum
The startup offers a machine learning infrastructure platform that provides a flexible operating system and virtualization interface for building and deploying machine learning and deep learning applications at scale. This technology enables enterprises to manage applications and hardware from a single terminal, resulting in increased productivity, reduced operational costs, and faster delivery times.
Armada
Armada provides an edge computing platform that integrates connectivity, ruggedized mobile data centers, and real-world AI applications to enable real-time data processing in remote environments. This technology addresses challenges in industries such as oil and gas, manufacturing, and logistics by enhancing safety, automating operations, and improving decision-making capabilities.
Funding: $50M+
Rough estimate of the amount of funding raised
DEEPX
Develops on-device AI semiconductor solutions, including custom NPUs, SoC ASICs, and specialized modules, optimized for low power consumption and high performance in applications like video analytics, security, and robotics. By enabling real-time AI processing with support for multiple models on a single chip, DEEPX addresses the challenges of latency, privacy, and network costs associated with cloud-based systems. Its scalable architecture and 259 patents ensure cost-competitive, silicon-proven products for global markets.
Funding: $100M+
Rough estimate of the amount of funding raised