1. Home
  2. AI Companies
  3. Graphcore
GR

Graphcore

About

Graphcore, a British semiconductor company and wholly owned subsidiary of SoftBank Group, develops specialized AI compute hardware centered on its Intelligence Processing Unit (IPU). The IPU represents a processor architecture specifically designed for machine intelligence workloads rather than general-purpose computing. The company built a complete AI compute stack spanning silicon design through datacenter infrastructure, including the Poplar software framework that sits atop the hardware. Graphcore brought the first Wafer-on-Wafer AI processor to market, a packaging approach that addresses the bandwidth and latency constraints inherent in traditional chip-to-chip interconnects for AI workloads.

The technical scope encompasses semiconductor engineering, processor design, and AI-specific optimizations across both hardware and software layers. The engineering team works on silicon design, wafer-scale integration technology, and the development of tools for AI model optimization. The software stack includes developer tools designed to extract performance from the IPU architecture, with ongoing work to optimize popular AI models for the platform. This systems-level approach attempts to address the throughput and efficiency bottlenecks that emerge when running large-scale machine learning workloads on conventional processor architectures.

Under CEO Nigel Toon's leadership, Graphcore operates with global presence and maintains teams of semiconductor, software, and AI specialists. The company's technology stack includes standard datacenter interfaces (PCIe, DDR, Ethernet) alongside proprietary elements like the IPU and Poplar software. The subsidiary structure under SoftBank provides backing for continued development of both the silicon and the software layers required to compete in AI compute infrastructure, where the trade-offs between custom silicon development costs and performance gains define commercial viability.

Open roles at Graphcore

Explore 133 open positions at Graphcore and find your next opportunity.

GR

Senior Bring-Up and Characterisation Engineer

Graphcore

Austin, Texas, United States (On-site)

$161.5K – $218.5K Yearly4d ago
GR

2026 Graduate Silicon Engineer

Graphcore

Cambridge, England, United Kingdom (On-site)

4d ago
GR

Staff Bring-Up and Characterisation Engineer

Graphcore

Austin, Texas, United States (On-site)

$198.1K – $268K Yearly4d ago
GR

System Test Manager

Graphcore

Bristol, England, United Kingdom (On-site)

4d ago
GR

2026 Graduate Silicon Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

4d ago
GR

Hardware Development Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

6d ago
GR

AI Performance Engineer

Graphcore

Milpitas, California, United States (On-site)

7d ago
GR

Principal Technical Program Manager

Graphcore

Bristol, England, United Kingdom (On-site)

1w ago
GR

Senior Thermal Engineer

Graphcore

Austin, Texas, United States (On-site)

1w ago
GR

Distinguished Engineer - Inference Serving Network and Storage

Graphcore

Austin, Texas, United States (On-site)

1w ago
GR

Pytorch Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

1w ago
GR

Senior Machine Learning Engineer (Large Systems)

Graphcore

Gdańsk, Pomeranian Voivodeship, Poland (On-site)

zł 260.4K – zł 352.2K Yearly1w ago
GR

Software Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

1w ago
GR

Director, Silicon Logical Design

Graphcore

Bristol, England, United Kingdom (On-site)

1w ago
GR

Senior Principal Test Framework Software Engineer

Graphcore

Austin, Texas, United States (On-site)

2w ago
GR

Principal SoC Architect

Graphcore

Bristol, England, United Kingdom (On-site)

2w ago
GR

Staff SoC Architect

Graphcore

Bristol, England, United Kingdom (On-site)

2w ago
GR

Research Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

2w ago
GR

Post Silicon Validation Engineer

Graphcore

Bristol, England, United Kingdom (On-site)

2w ago

Similar companies

CO

CoreWeave

CoreWeave operates specialized cloud infrastructure purpose-built for AI workloads, with data centers across the US and Europe delivering GPU compute for large language model training and inference at scale. Founded in 2017 as Atlantic Crypto, a cryptocurrency mining operation, the company executed a complete strategic pivot to AI infrastructure - rebuilding from first principles rather than retrofitting existing cloud architectures. The platform runs on Kubernetes-based orchestration designed specifically for AI workloads, coupled with custom storage solutions engineered to handle the I/O patterns and throughput requirements of model training and deployment pipelines. The technical stack centers on NVIDIA GPUs with orchestration built in Go, Python, and C++ on Linux, instrumented with Prometheus, Grafana, and OpenTelemetry for observability across distributed systems. Rather than adapting general-purpose cloud tooling, CoreWeave's infrastructure treats GPU compute density, inter-node bandwidth, and storage parallelism as primary design constraints. This systems-level focus reflects a team drawn from infrastructure engineering and quantitative trading backgrounds - disciplines where latency budgets and resource utilization directly determine feasibility. CoreWeave serves AI labs, enterprises, and startups requiring production-scale inference and training capacity. The company's recognition on the TIME100 most influential companies list signals market adoption of specialized AI infrastructure as distinct from traditional cloud providers. For engineers, the environment offers direct exposure to the operational realities of running GPU clusters at scale: thermal management, network topology for distributed training, failure modes in multi-tenant GPU environments, and the cost-performance trade-offs inherent in serving latency-sensitive inference workloads alongside batch training jobs.

436 jobs
MA

Mistral AI

Mistral AI is a French AI company founded in April 2023 by Arthur Mensch, Guillaume Lample, and Timothée Lacroix - researchers with prior affiliations at Google DeepMind and Meta and academic roots at École Polytechnique. The company develops and releases open-weight, state-of-the-art generative AI models positioned as alternatives to proprietary solutions, with a focus on democratizing access to frontier AI technology. Their core approach centers on open, transparent model development that enables developers, enterprises, and institutions to build applications while maintaining control over their data and deployments. The company's primary product line consists of open-weight generative AI models released publicly, which Mistral claims rival proprietary solutions in capability. Their technical domains span generative AI model training, with particular emphasis on open-weight architectures, AI transparency, and bias mitigation. The founding mission explicitly opposes what the company characterizes as emerging opacity and centralization in AI systems, positioning their open-weight approach as a structural alternative to closed, proprietary models. Mistral AI's operational model emphasizes community-backed development and targets a broad user base spanning individual developers, enterprise deployments, and institutional applications across global markets. The company's cultural positioning centers on maintaining user control over inference infrastructure and data pipelines, combating censorship in model outputs, and providing an alternative to concentrated control of frontier AI capabilities. While specific scale metrics around model performance, deployment volumes, or operational characteristics are not publicly detailed, the company claims to have achieved state-of-the-art results in their released model family.

212 jobs
CE

Cerebras

Cerebras Systems designs and manufactures wafer-scale AI chips that consolidate the compute capacity of dozens of GPUs into a single device. Founded in 2015, the company's core architecture is 56 times larger than standard GPUs, addressing the operational complexity of distributed training and inference by offering programmability equivalent to a single-device system while delivering multi-GPU performance. This approach collapses the network bottlenecks and synchronization overhead inherent in GPU clusters, enabling users to run large-scale ML workloads without orchestrating hundreds of accelerators. The company's technical stack spans the full systems hierarchy: custom silicon (wafer-scale chip architecture), compiler infrastructure (MLIR, LLVM IR, and their proprietary CSL language), runtime orchestration (Kubernetes), and deployment tooling. Engineering work touches computer architecture, deep learning kernels, systems software for hardware programmability, and inference serving at scale. Recent partnerships include work with OpenAI on inference deployment, alongside engagements with national laboratories, global enterprises, and healthcare systems requiring high-throughput ML serving. Cerebras positions its hardware for both training and inference workloads, with claimed industry-leading speeds stemming from on-chip interconnect bandwidth and elimination of multi-chip communication latency. The architecture trades traditional data center modularity for integrated performance - relevant for workloads bottlenecked by cross-device synchronization or where cost-per-inference and tail latency matter more than incremental horizontal scaling. Development infrastructure includes C++, Python, Go, and Zig across the stack, with CI/CD through GitHub Actions and Jenkins.

135 jobs
RU

Runpod

RunPod operates an end-to-end AI infrastructure platform focused on GPU compute provisioning for model training, inference, and distributed agent orchestration. The platform serves over 500,000 developers, spanning solo practitioners to enterprise teams deploying at scale. Core infrastructure handles compute allocation, orchestration complexity, and operational overhead - positioning itself as accessible infrastructure rather than requiring deep systems expertise from users. The technical stack centers on Go, Python, and TypeScript with containerization through Docker and Kubernetes orchestration on Linux. Engineering domains span distributed systems, GPU compute scheduling, and developer tooling designed to abstract provisioning and scaling mechanics. The company emphasizes reducing operational friction: developers interact with compute resources without managing underlying cluster complexity or infrastructure provisioning bottlenecks. RunPod maintains a remote-first structure with team distribution across the U.S., Canada, Europe, and India. The platform's design reflects a systems-first approach to making GPU compute economically viable and operationally manageable - targeting workloads where cost, reliability, and time-to-deployment constrain AI development cycles.

26 jobs
VA

Vast.ai

Vast.ai operates a peer-to-peer GPU marketplace connecting over 10,000 GPUs across 40 data centers with users requiring compute for training, fine-tuning, and inference workloads. The platform aggregates capacity from data centers and individual providers running Vast's hosting software, offering on-demand, interruptible, and auction-based pricing models that price 3-5x below traditional cloud providers. Instance deployment occurs in seconds, with the marketplace enabling direct comparison of price-performance across heterogeneous hardware. The architecture surfaces a pricing-availability trade-off inherent to peer-to-peer models: cost savings derive from utilizing underutilized capacity, but availability and reliability vary by provider. Interruptible instances present the sharpest cost-performance point but require fault-tolerant workloads and checkpointing discipline. The platform supports standard ML frameworks (PyTorch, TensorFlow) and containerized deployments via Docker. Enterprise offerings provide dedicated clusters with SLAs, SOC 2 Type I certification, and access to ISO 27001 certified facilities, trading marketplace economics for operational predictability. The technical stack spans Python and C++ for core platform services, PostgreSQL for marketplace state, Redis for coordination, and Terraform for infrastructure provisioning. CUDA support is foundational for GPU workloads. The system must handle heterogeneous provider configurations, node churn, and pricing dynamics across thousands of GPUs while maintaining search and allocation latency suitable for rapid instance provisioning. Founded in 2018, the company positions itself as infrastructure for cost-sensitive training and inference at scale.

18 jobs