About

xAI, founded by Elon Musk in 2023, builds AI systems designed to advance scientific discovery and gain deeper understanding of the universe. The company operates across offices in Palo Alto, Seattle, San Francisco, Tennessee, and London, with a technical infrastructure spanning Python, Rust, JAX, and Kubernetes for model development and deployment, alongside TypeScript, React, and WebAssembly for interface layers. The engineering stack emphasizes RoCEv2 and InfiniBand for networking in distributed training and inference workloads.

The company's flagship product is Grok, a conversational AI modeled after the Hitchhiker's Guide to the Galaxy, providing real-time information access integrated with the X platform. Development follows first-principles reasoning with rapid iteration cycles, focusing on system bottlenecks in latency, throughput, and reliability rather than incremental feature additions. The technical approach centers on large language model architectures optimized for both scientific reasoning tasks and production conversational inference at scale.

xAI's engineering culture prioritizes operational complexity trade-offs inherent in deploying large models - managing tail latency in multi-tenant inference serving, balancing cost against throughput requirements, and addressing failure modes in real-time information retrieval systems. The team composition spans researchers and engineers working on problems at the intersection of AI capabilities research and production system reliability, with infrastructure supporting both research experimentation and user-facing deployment.

Open roles at xAI

Explore 222 open positions at xAI and find your next opportunity.

XA

Senior Analyst - Safety Operations (CSE)

xAI

Palo Alto, California, United States (On-site)

$44 – $63 Hourly2w ago
XA

AI Tutor - Marathi

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Hebrew

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Tamil

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Swedish

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Urdu

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - French

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Telugu

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Thai

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Norwegian

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Turkish

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Hindi

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Indonesian

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Spanish

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Italian

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Vietnamese

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

Safety Operations Senior Analyst

xAI

Bastrop, Texas, United States (On-site)

2w ago
XA

AI Tutor - Finnish

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Japanese

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago
XA

AI Tutor - Korean

xAI

Worldwide (Remote)

$35 – $45 Hourly2w ago

Similar companies

OP

OpenAI

OpenAI develops and deploys generative transformer models at scale, operating production systems that serve millions through ChatGPT, GPT model APIs, and the OpenAI API. The technical challenge spans the full stack: research engineering for novel model architectures, safety engineering for alignment and robustness, and production infrastructure for API deployment at scale. Teams work across research, product engineering, and operations, with work organized around both advancing model capabilities and maintaining reliability for deployed systems serving substantial user traffic. The core technical domains include model development for the GPT series, API infrastructure to support downstream applications, and safety research focused on making AGI beneficial. Engineering work involves trade-offs between model capability, inference cost, latency characteristics, and safety constraints. Research teams collaborate with product and engineering functions to move from experimental systems to production deployment, requiring expertise in distributed systems, model optimization, and operational complexity at scale. The company operates from San Francisco with international presence, positioning work as a global effort toward artificial general intelligence. Cross-functional teams include researchers, engineers, and operations staff working on problems ranging from foundational research to production reliability. The technical culture emphasizes rigorous safety practices alongside advancement of capabilities, with autonomy and ownership distributed across teams working on distinct components of the research-to-deployment pipeline.

741 jobs
HA

Harvey

Harvey operates a domain-specific AI platform serving hundreds of law firms and Fortune 500 companies worldwide. Founded in 2022, the company integrates cutting-edge AI models with deep industry expertise to handle contract analysis, legal research, and document review across law, tax, and finance verticals. The platform addresses the complexity of professional service workflows where domain knowledge and reliability requirements constrain general-purpose AI deployment. The technical stack spans multi-cloud infrastructure (Azure, GCP, AWS) with Kubernetes orchestration and infrastructure-as-code tooling (Terraform, Pulumi, CloudFormation). Observability runs through Datadog, Sentry, and PagerDuty. Primary development occurs in Python and Go with Redis for caching and state management. The platform's architecture prioritizes security and reliability - critical operational constraints given the sensitive nature of legal and financial documents and the professional liability exposure of law firm customers. The team combines former lawyers, AI researchers, and engineers, reflecting the dual constraint of building production AI systems while maintaining the domain accuracy required by professional service providers. Customer engagement centers on workflow integration: understanding how professionals actually work and building tools that fit existing processes rather than forcing adoption of new patterns. This approach addresses the adoption friction typical in professional services, where productivity tools must demonstrate clear value against high switching costs and risk-averse user populations.

196 jobs
VE

Vertiv

Vertiv operates critical digital infrastructure at global scale, delivering end-to-end systems that power and cool data centers, communication networks, and commercial facilities. The company's technical scope spans grid-to-chip power chains, thermal management, and intelligent monitoring - infrastructure that determines operational availability and performance characteristics for compute workloads from edge deployments to hyperscale cloud environments. With decades of domain expertise, Vertiv addresses the operational bottlenecks inherent in maintaining continuous uptime for mission-critical applications. The product portfolio reflects infrastructure constraints across the stack: critical power solutions that maintain grid-to-chip continuity, adaptive cooling systems calibrated for varying thermal loads, and liquid cooling technologies designed specifically for high-density compute environments where traditional air cooling becomes a throughput limiter. Modular prefabricated data centers enable deployment at speed, while advanced battery energy storage systems provide backup power with different trade-offs than traditional UPS architectures. Intelligent monitoring and management systems surface operational visibility across these integrated components. Vertiv serves customers ranging from hyperscale cloud providers managing efficiency at massive scale to local telecommunications networks with different reliability and cost constraints. The company positions its systems around operational excellence and business continuity - measurable outcomes in environments where infrastructure failures directly impact application availability. Digital services and expert support complement hardware deployment, addressing the operational complexity of maintaining critical infrastructure across geographically distributed sites. Under CEO Giordano Albertazzi, the company maintains a hardware-first approach while incorporating software monitoring and management capabilities, with stated emphasis on sustainability goals alongside traditional reliability metrics.

140 jobs
LA

LangChain

LangChain operates an engineering platform and open source frameworks for building, testing, and deploying AI agents. The core offering comprises LangChain and LangGraph - open source frameworks providing pre-built architectures and access to 1,000+ integrations - alongside LangSmith, a commercial platform for observability, evaluation, and deployment of LLM systems. The frameworks see over 90 million combined downloads per month and are used by millions of developers worldwide, with named deployments at Replit, Clay, Cloudflare, Harvey, Rippling, Vanta, Workday, LinkedIn, and Coinbase. The technical stack addresses the production bottlenecks of agent engineering: reliability through comprehensive observability, evaluation tooling to surface failure modes before deployment, and deployment infrastructure to move from prototype to production. LangSmith's platform provides the operational layer for teams moving LLM systems into production environments, while the open source frameworks prioritize development velocity through pre-built components and extensive integration coverage. The architecture allows granular control over agent behavior while reducing the complexity of integrating external services and managing LLM system reliability at scale. LangChain serves both major enterprises and startups building AI agents, with technical domains spanning agent engineering, LLM systems, observability, evaluation, and developer tooling. The company is led by CEO Harrison Chase and maintains a US headquarters, with a worldwide developer base. The dual model of open source frameworks and commercial platform reflects a focus on production-readiness and operational support for teams deploying agents at scale.

114 jobs
CO

Cohere

Cohere builds enterprise-focused foundational models designed for production deployment with emphasis on security, privacy, and operational trust. Founded in 2019 in Toronto, the company has raised nearly $1 billion and scaled to hundreds of employees worldwide. The technical focus spans semantic search, content generation, and customer experience applications - domains where model reliability and data governance are non-negotiable constraints for enterprise adoption. The company's architecture decisions reflect production realities over research novelty. Models are architected for deployment into regulated environments where data residency, access controls, and audit trails matter as much as accuracy metrics. This positioning addresses the gap between frontier model capabilities and enterprise operational requirements: latency SLAs, cost predictability, and compliance frameworks that prevent many organizations from operationalizing public AI APIs. Cohere Labs has published over 100 papers and built a research community of 4,500+ researchers, signaling ongoing investment in foundational work rather than pure application-layer focus. The team composition skews heavily toward researchers and engineers from academic backgrounds, which maps to the technical challenge space - building models that balance performance, safety constraints, and deployment flexibility across varied enterprise infrastructure.

106 jobs
BA

Baseten

Baseten builds AI infrastructure for production deployment and scaling of models, with work spanning kernel-level optimization for inference performance through developer tooling. The platform ships daily, measuring success by real-world impact of AI products running on it rather than vanity metrics. Engineers embed directly with customers to surface operational bottlenecks, then optimize obsessively - work ranges from TensorRT-LLM and CUDA kernel tuning to building developer tools that reduce deployment friction. The stack centers on inference at scale: TensorRT-LLM and PyTorch for model execution, NVIDIA Triton Inference Server for serving, Kubernetes (EKS) with Karpenter for autoscaling, and Knative for event-driven workloads on AWS EC2. Infrastructure decisions prioritize shipping velocity over process - small teams with real ownership iterate rapidly on production reliability, latency (including tail behavior), and cost efficiency. Docker containerization and PostgreSQL round out core operational dependencies. The team is internationally distributed, composed of engineers and designers who take craft seriously without performative posturing. Customer-embedded engineering informs both platform architecture and developer experience tradeoffs, creating tight feedback loops between deployment reality and infrastructure evolution. From founding, the approach has centered on hands-on problem solving and rapid iteration rather than abstraction layers that delay production learning.

69 jobs