MO

About

Modal operates a serverless compute platform designed to minimize infrastructure friction for ML inference, fine-tuning, and batch workloads. The platform provides instant GPU access with usage-based pricing, targeting teams that need to ship compute-intensive applications without managing scheduling, container orchestration, or resource allocation. The architecture is built on custom infrastructure components - an in-house file system, container runtime, scheduler, and image builder - optimized for the latency and throughput characteristics of AI workloads.

The technical stack spans Python, Rust, and Go at the systems level, with PyTorch, CUDA, vLLM, and TensorRT support for ML frameworks. This reflects prioritization of both developer ergonomics (Python interface) and low-level performance (Rust/Go for runtime components). The custom infrastructure signals investment in controlling the full vertical - from container initialization through GPU scheduling - rather than composing existing orchestration layers.

The team operates across New York, Stockholm, and San Francisco, and includes creators of open-source projects like Seaborn and Luigi, alongside academic researchers and engineers with experience building production systems. The platform positions itself around developer experience as a core constraint, with infrastructure complexity abstracted to reduce operational overhead for data and AI teams.

Open roles at Modal

Explore 25 open positions at Modal and find your next opportunity.

MO

Member of Technical Staff - Agent DX Research

Modal

New York, United States (On-site)

$150K – $350K Yearly1mo ago
MO

Member of Technical Staff - ML Training Systems

Modal

New York, United States (On-site)

$150K – $350K Yearly1mo ago
MO

Forward Deployed Engineer - ML

Modal

Stockholm, Sweden (On-site)

1mo ago
MO

VP Finance

Modal

New York, United States (On-site)

$300K – $350K Yearly2mo ago
MO

Forward Deployed ML Engineer

Modal

New York, United States (On-site)

$180K – $250K Yearly2mo ago
MO

Solutions Architect

Modal

San Francisco, California, United States (Hybrid)

$200K – $280K Yearly2mo ago
MO

Systems Engineering Manager

Modal

New York, United States (On-site)

$250K – $350K Yearly2mo ago
MO

Business Operations Manager

Modal

New York, United States (On-site)

$125K – $175K Yearly2mo ago
MO

People & Talent Lead

Modal

Stockholm, Stockholm, Sweden (On-site)

2mo ago
MO

Member of Technical Staff - Systems

Modal

Stockholm, Stockholm, Sweden (On-site)

3mo ago
MO

Controller

Modal

New York, United States (On-site)

$220K – $235K Yearly3mo ago
MO

Member of Technical Staff - Python SDK

Modal

New York, United States (On-site)

$150K – $270K Yearly3mo ago
MO

Member of Technical Staff - Product (Frontend)

Modal

New York, United States (On-site)

$150K – $270K Yearly3mo ago
MO

Member of Technical Staff - ML Performance

Modal

New York, United States (On-site)

$150K – $270K Yearly3mo ago
MO

Founding GTM Talent Partner

Modal

New York, United States (On-site)

$150K – $195K Yearly3mo ago
MO

Security Engineer

Modal

New York, United States (On-site)

$150K – $270K Yearly3mo ago
MO

Developer Relations Engineer

Modal

San Francisco, California, United States (On-site)

$175K – $275K Yearly3mo ago
MO

Member of Technical Staff - Systems

Modal

New York, United States (On-site)

$150K – $270K Yearly3mo ago
MO

Forward Deployed Engineer - Systems

Modal

San Francisco, California, United States or Remote (California, United States + 2 more)

$180K – $240K Yearly3mo ago
MO

Technical Content Marketing

Modal

New York, United States (On-site)

$130K – $250K Yearly3mo ago

Similar companies

DE

Decagon

Decagon builds a conversational AI platform designed to replace or augment legacy customer support systems by deploying intelligent AI agents across chat, email, and voice channels. The company positions its technology as infrastructure for delivering concierge-level customer experiences at scale, targeting brands looking to support, onboard, and retain customers without proportional headcount growth. Led by CEO Jesse Zhang and founded by serial entrepreneurs, Decagon operates from the US and focuses on addressing the operational constraints of traditional customer support systems. The platform's core technical approach centers on Agent Operating Procedures (AOPs), a natural-language-to-code compilation system that allows non-technical users to define agent behavior while preserving technical team control over guardrails, integrations, and versioning. This design addresses a common trade-off in AI tooling: enabling rapid iteration by domain experts without sacrificing reliability controls or introducing configuration drift. The agent orchestration layer spans multiple channels and claims to amplify CX team impact by 10x, though specific benchmarks around latency, accuracy, or failure rate are not publicly detailed. Decagon's technical domains span conversational AI, natural language processing, multichannel messaging infrastructure, and automation systems. The platform emphasizes runtime guardrails and version management as first-class concerns, reflecting a systems-oriented approach to production deployment. The company claims to deliver always-on, personalized service, positioning its agents as operational infrastructure rather than experimental tooling. For engineers evaluating opportunities, the technical challenges likely involve scaling context-rich, stateful interactions across channels while maintaining consistency, handling edge cases in natural language understanding, and building abstraction layers that balance expressiveness with safety.

89 jobs
MA

Magic

Magic operates 8,000 NVIDIA H100s on Google Cloud, training frontier code models designed to automate software engineering and AI research itself. The company has raised $515 million from Nat Friedman, Daniel Gross, CapitalG, and Sequoia to pursue direct AGI development through code generation - treating automated AI research as the primary bottleneck rather than incremental developer tooling. Technical focus spans large-scale pre-training, domain-specific reinforcement learning, ultra-long context windows, and inference-time compute scaling. The company's research program centers on fundamental problems in automating software engineering at scale, not incremental productivity improvements. Context window extension and inference-time compute are treated as first-class constraints rather than auxiliary features. Co-founded by Eric Steinberger and Sebastian De Ro, the team remains small and emphasizes ownership over execution - engineers and researchers work on meaningful problem subsets rather than predetermined roadmaps. Infrastructure operates at production scale: the H100 cluster represents committed capital toward training runs that matter, not research prototypes. The operational model assumes that code generation quality and AI research automation are the direct path to AGI, making software engineering the domain where model capabilities and safety research converge. Google Cloud provides the substrate, but the company owns its GPU allocation outright.

17 jobs
AM