About

Harvey operates a domain-specific AI platform serving hundreds of law firms and Fortune 500 companies worldwide. Founded in 2022, the company integrates cutting-edge AI models with deep industry expertise to handle contract analysis, legal research, and document review across law, tax, and finance verticals. The platform addresses the complexity of professional service workflows where domain knowledge and reliability requirements constrain general-purpose AI deployment.

The technical stack spans multi-cloud infrastructure (Azure, GCP, AWS) with Kubernetes orchestration and infrastructure-as-code tooling (Terraform, Pulumi, CloudFormation). Observability runs through Datadog, Sentry, and PagerDuty. Primary development occurs in Python and Go with Redis for caching and state management. The platform's architecture prioritizes security and reliability - critical operational constraints given the sensitive nature of legal and financial documents and the professional liability exposure of law firm customers.

The team combines former lawyers, AI researchers, and engineers, reflecting the dual constraint of building production AI systems while maintaining the domain accuracy required by professional service providers. Customer engagement centers on workflow integration: understanding how professionals actually work and building tools that fit existing processes rather than forcing adoption of new patterns. This approach addresses the adoption friction typical in professional services, where productivity tools must demonstrate clear value against high switching costs and risk-averse user populations.

Open roles at Harvey

Explore 162 open positions at Harvey and find your next opportunity.

HA

Enterprise Sales Manager, Paris

Harvey

France or Remote (France)

3mo ago
HA

Senior Product Marketing Manager

Harvey

Manhattan, New York, New York, United States (Hybrid)

$145K – $195K Yearly3mo ago

Similar companies

XA

xAI

xAI, founded by Elon Musk in 2023, builds AI systems designed to advance scientific discovery and gain deeper understanding of the universe. The company operates across offices in Palo Alto, Seattle, San Francisco, Tennessee, and London, with a technical infrastructure spanning Python, Rust, JAX, and Kubernetes for model development and deployment, alongside TypeScript, React, and WebAssembly for interface layers. The engineering stack emphasizes RoCEv2 and InfiniBand for networking in distributed training and inference workloads. The company's flagship product is Grok, a conversational AI modeled after the Hitchhiker's Guide to the Galaxy, providing real-time information access integrated with the X platform. Development follows first-principles reasoning with rapid iteration cycles, focusing on system bottlenecks in latency, throughput, and reliability rather than incremental feature additions. The technical approach centers on large language model architectures optimized for both scientific reasoning tasks and production conversational inference at scale. xAI's engineering culture prioritizes operational complexity trade-offs inherent in deploying large models - managing tail latency in multi-tenant inference serving, balancing cost against throughput requirements, and addressing failure modes in real-time information retrieval systems. The team composition spans researchers and engineers working on problems at the intersection of AI capabilities research and production system reliability, with infrastructure supporting both research experimentation and user-facing deployment.

360 jobs
VE

Vertiv

Vertiv operates critical digital infrastructure at global scale, delivering end-to-end systems that power and cool data centers, communication networks, and commercial facilities. The company's technical scope spans grid-to-chip power chains, thermal management, and intelligent monitoring - infrastructure that determines operational availability and performance characteristics for compute workloads from edge deployments to hyperscale cloud environments. With decades of domain expertise, Vertiv addresses the operational bottlenecks inherent in maintaining continuous uptime for mission-critical applications. The product portfolio reflects infrastructure constraints across the stack: critical power solutions that maintain grid-to-chip continuity, adaptive cooling systems calibrated for varying thermal loads, and liquid cooling technologies designed specifically for high-density compute environments where traditional air cooling becomes a throughput limiter. Modular prefabricated data centers enable deployment at speed, while advanced battery energy storage systems provide backup power with different trade-offs than traditional UPS architectures. Intelligent monitoring and management systems surface operational visibility across these integrated components. Vertiv serves customers ranging from hyperscale cloud providers managing efficiency at massive scale to local telecommunications networks with different reliability and cost constraints. The company positions its systems around operational excellence and business continuity - measurable outcomes in environments where infrastructure failures directly impact application availability. Digital services and expert support complement hardware deployment, addressing the operational complexity of maintaining critical infrastructure across geographically distributed sites. Under CEO Giordano Albertazzi, the company maintains a hardware-first approach while incorporating software monitoring and management capabilities, with stated emphasis on sustainability goals alongside traditional reliability metrics.

140 jobs
CO

Cohere

Cohere builds enterprise-focused foundational models designed for production deployment with emphasis on security, privacy, and operational trust. Founded in 2019 in Toronto, the company has raised nearly $1 billion and scaled to hundreds of employees worldwide. The technical focus spans semantic search, content generation, and customer experience applications - domains where model reliability and data governance are non-negotiable constraints for enterprise adoption. The company's architecture decisions reflect production realities over research novelty. Models are architected for deployment into regulated environments where data residency, access controls, and audit trails matter as much as accuracy metrics. This positioning addresses the gap between frontier model capabilities and enterprise operational requirements: latency SLAs, cost predictability, and compliance frameworks that prevent many organizations from operationalizing public AI APIs. Cohere Labs has published over 100 papers and built a research community of 4,500+ researchers, signaling ongoing investment in foundational work rather than pure application-layer focus. The team composition skews heavily toward researchers and engineers from academic backgrounds, which maps to the technical challenge space - building models that balance performance, safety constraints, and deployment flexibility across varied enterprise infrastructure.

106 jobs
PI

Pinecone

Pinecone operates a fully managed vector database service designed for production AI applications requiring storage and retrieval of high-dimensional embeddings. The system handles vector search at scale across recommendation systems, semantic search, and related ML-backed services. Founded by Edo Liberty, formerly a research director at AWS with prior experience building custom vector search systems at large scale, the company is credited with establishing the vector database category as a distinct infrastructure layer. The technical stack centers on systems languages - Rust, Go, C++, and Python - with RocksDB as the storage engine and Kubernetes orchestration across AWS, GCP, and Azure. This architecture targets the operational complexity of managing embedding indices, query latency, and throughput at production scale, abstracting infrastructure decisions from engineering teams deploying AI features. The platform serves thousands of companies, positioning itself on ease of deployment and reduced time-to-production for vector-backed applications. The founding principle emphasizes accessibility for engineering teams of varying sizes, evolving the managed service model to minimize operational overhead in running vector workloads. Core focus areas include retrieval performance, reliability under production load, and cost-efficiency trade-offs inherent to high-dimensional search systems.

9 jobs