About

Magic operates 8,000 NVIDIA H100s on Google Cloud, training frontier code models designed to automate software engineering and AI research itself. The company has raised $515 million from Nat Friedman, Daniel Gross, CapitalG, and Sequoia to pursue direct AGI development through code generation - treating automated AI research as the primary bottleneck rather than incremental developer tooling. Technical focus spans large-scale pre-training, domain-specific reinforcement learning, ultra-long context windows, and inference-time compute scaling.

The company's research program centers on fundamental problems in automating software engineering at scale, not incremental productivity improvements. Context window extension and inference-time compute are treated as first-class constraints rather than auxiliary features. Co-founded by Eric Steinberger and Sebastian De Ro, the team remains small and emphasizes ownership over execution - engineers and researchers work on meaningful problem subsets rather than predetermined roadmaps.

Infrastructure operates at production scale: the H100 cluster represents committed capital toward training runs that matter, not research prototypes. The operational model assumes that code generation quality and AI research automation are the direct path to AGI, making software engineering the domain where model capabilities and safety research converge. Google Cloud provides the substrate, but the company owns its GPU allocation outright.

Open roles at Magic

Explore 12 open positions at Magic and find your next opportunity.

MA

IT Support Engineer

Magic

San Francisco, California, United States (On-site)

$170K – $230K Yearly1mo ago
MA

Software Engineer

Magic

San Francisco, California, United States (On-site)

$200K – $550K Yearly1mo ago
MA

Member of Technical Staff, Inference & RL Systems

Magic

San Francisco, California, United States (On-site)

$225K – $550K Yearly1mo ago
MA

Member of Technical Staff, Pre-training Systems

Magic

San Francisco, California, United States (On-site)

$225K – $550K Yearly1mo ago
MA

Member of Technical Staff, Pre-training Data

Magic

San Francisco, California, United States (On-site)

$200K – $550K Yearly1mo ago
MA

Technical Sourcer

Magic

San Francisco, California, United States (On-site)

$150K – $260K Yearly3mo ago
MA

Software Engineer - Post-training Data

Magic

San Francisco, California, United States (On-site)

$200K – $550K Yearly3mo ago
MA

Software Engineer - Supercomputing Platform & Infrastructure

Magic

San Francisco, California, United States (On-site)

$200K – $550K Yearly3mo ago
MA

Kernel Engineer

Magic

San Francisco, California, United States (On-site)

$225K – $550K Yearly3mo ago
MA

<insert-job-you-excel-at/>

Magic

San Francisco, California, United States or Remote (United States)

$100K – $550K Yearly3mo ago
MA

Research Engineer

Magic

San Francisco, California, United States (On-site)

$225K – $550K Yearly3mo ago
MA

Software Engineer - Product

Magic

San Francisco, California, United States (On-site)

$200K – $550K Yearly3mo ago

Similar companies

EL

ElevenLabs

ElevenLabs is an AI audio research and deployment company building voice AI systems that serve millions of developers, creators, and enterprises. The company's technical focus spans speech synthesis, voice cloning, multilingual voice models, and conversational AI agents. Their models support over 70 languages, with core capabilities in text-to-speech, sound effects generation, and voice agent deployment. The company is backed by Andreessen Horowitz, Sequoia, and other investors. The platform consists of three main products: ultra-realistic AI voices designed for clarity, expressiveness, and multilingual support; an Agents Platform that enables teams to deploy voice agents capable of listening, talking, and acting; and a Creative Platform focused on content localization, storytelling, and accessibility improvements. Primary technical domains include speech synthesis systems, voice cloning infrastructure, and conversational agent platforms built on Python and TypeScript. ElevenLabs serves businesses ranging from early-stage startups to large enterprises across multiple verticals: customer support, sales automation, education, video production, publishing, and accessibility applications. Named use cases include reading articles, voice-over generation, voice restoration for individuals with disabilities, and building intelligent agents for support, sales, and education workflows. The company's operational model emphasizes both research and production deployment, with infrastructure supporting content localization and audio-based applications at scale.

124 jobs
RE

Replit

Replit operates a web-based code editor and multiplayer computing environment used by millions for collaborative software development. The platform eliminates traditional barriers to application creation through natural language interfaces, allowing users to build applications without conventional development workflows - demonstrated by architectural decisions like removing the save button from their editor. The multiplayer environment serves as infrastructure for experimentation, sharing, and collaborative growth at scale. The company measures success by the number of people empowered to create software rather than vanity metrics, reflecting a systems-level focus on removing bottlenecks in developer onboarding and productivity. Technical decisions prioritize shipping velocity and operational autonomy: the culture emphasizes extreme ownership, radical bets, and bias toward action. Engineers operate with the latitude to pursue emergent ideas and question established patterns when friction appears in the development loop. The platform's architecture supports collaborative coding workflows at scale, handling millions of concurrent users across a shared computing environment. This requires managing trade-offs between multi-tenancy constraints, latency in collaborative editing, and operational complexity of maintaining compute resources for distributed development sessions. The technical focus centers on developer tools, web-based editing infrastructure, and the reliability challenges of real-time collaborative computing.

76 jobs
MO

Modal

Modal operates a serverless compute platform designed to minimize infrastructure friction for ML inference, fine-tuning, and batch workloads. The platform provides instant GPU access with usage-based pricing, targeting teams that need to ship compute-intensive applications without managing scheduling, container orchestration, or resource allocation. The architecture is built on custom infrastructure components - an in-house file system, container runtime, scheduler, and image builder - optimized for the latency and throughput characteristics of AI workloads. The technical stack spans Python, Rust, and Go at the systems level, with PyTorch, CUDA, vLLM, and TensorRT support for ML frameworks. This reflects prioritization of both developer ergonomics (Python interface) and low-level performance (Rust/Go for runtime components). The custom infrastructure signals investment in controlling the full vertical - from container initialization through GPU scheduling - rather than composing existing orchestration layers. The team operates across New York, Stockholm, and San Francisco, and includes creators of open-source projects like Seaborn and Luigi, alongside academic researchers and engineers with experience building production systems. The platform positions itself around developer experience as a core constraint, with infrastructure complexity abstracted to reduce operational overhead for data and AI teams.

28 jobs