- Home
- AI Companies
- Harmonic
Harmonic
Similar companies
OpenAI
OpenAI develops and deploys generative transformer models at scale, operating production systems that serve millions through ChatGPT, GPT model APIs, and the OpenAI API. The technical challenge spans the full stack: research engineering for novel model architectures, safety engineering for alignment and robustness, and production infrastructure for API deployment at scale. Teams work across research, product engineering, and operations, with work organized around both advancing model capabilities and maintaining reliability for deployed systems serving substantial user traffic. The core technical domains include model development for the GPT series, API infrastructure to support downstream applications, and safety research focused on making AGI beneficial. Engineering work involves trade-offs between model capability, inference cost, latency characteristics, and safety constraints. Research teams collaborate with product and engineering functions to move from experimental systems to production deployment, requiring expertise in distributed systems, model optimization, and operational complexity at scale. The company operates from San Francisco with international presence, positioning work as a global effort toward artificial general intelligence. Cross-functional teams include researchers, engineers, and operations staff working on problems ranging from foundational research to production reliability. The technical culture emphasizes rigorous safety practices alongside advancement of capabilities, with autonomy and ownership distributed across teams working on distinct components of the research-to-deployment pipeline.
Crusoe
Crusoe designs, builds, and operates purpose-built data centers and cloud computing infrastructure powered by renewable energy sources including wind, solar, geothermal, and hydropower. Founded in 2018 by Chase Lochmiller and Cully Cavness, the company operates gigawatt-scale data center campuses and has raised billions in funding. The infrastructure supports AI workloads through partnerships with NVIDIA and AMD, offering GPU-backed cloud services focused on the trade-off between computational scale and energy sustainability. The company's technical stack spans cluster orchestration (Kubernetes, Slurm), infrastructure automation (Terraform, Ansible, Puppet), and distributed storage systems (Ceph, GlusterFS, OpenEBS). Development work involves Python, Go, Java, and C, with infrastructure built on Linux, NVMe storage, and RDMA networking to support high-throughput AI training and inference workloads. The vertical integration approach extends from data center construction through hardware partnerships to cloud platform operations. Crusoe evolved from early operations converting wasted natural gas from oil fields into computing power for bitcoin mining. The current focus is AI infrastructure delivery, where the energy-first approach addresses the operational constraint of power availability at scale - a bottleneck increasingly relevant as model size and inference volume grow. The cloud platform enables organizations to deploy AI solutions with access to GPU capacity backed by renewable energy sources, though specific performance characteristics, availability zones, and pricing models are not publicly detailed in standard materials.
CoreWeave
CoreWeave operates specialized cloud infrastructure purpose-built for AI workloads, with data centers across the US and Europe delivering GPU compute for large language model training and inference at scale. Founded in 2017 as Atlantic Crypto, a cryptocurrency mining operation, the company executed a complete strategic pivot to AI infrastructure - rebuilding from first principles rather than retrofitting existing cloud architectures. The platform runs on Kubernetes-based orchestration designed specifically for AI workloads, coupled with custom storage solutions engineered to handle the I/O patterns and throughput requirements of model training and deployment pipelines. The technical stack centers on NVIDIA GPUs with orchestration built in Go, Python, and C++ on Linux, instrumented with Prometheus, Grafana, and OpenTelemetry for observability across distributed systems. Rather than adapting general-purpose cloud tooling, CoreWeave's infrastructure treats GPU compute density, inter-node bandwidth, and storage parallelism as primary design constraints. This systems-level focus reflects a team drawn from infrastructure engineering and quantitative trading backgrounds - disciplines where latency budgets and resource utilization directly determine feasibility. CoreWeave serves AI labs, enterprises, and startups requiring production-scale inference and training capacity. The company's recognition on the TIME100 most influential companies list signals market adoption of specialized AI infrastructure as distinct from traditional cloud providers. For engineers, the environment offers direct exposure to the operational realities of running GPU clusters at scale: thermal management, network topology for distributed training, failure modes in multi-tenant GPU environments, and the cost-performance trade-offs inherent in serving latency-sensitive inference workloads alongside batch training jobs.
Cohere
Cohere builds enterprise-focused foundational models designed for production deployment with emphasis on security, privacy, and operational trust. Founded in 2019 in Toronto, the company has raised nearly $1 billion and scaled to hundreds of employees worldwide. The technical focus spans semantic search, content generation, and customer experience applications - domains where model reliability and data governance are non-negotiable constraints for enterprise adoption. The company's architecture decisions reflect production realities over research novelty. Models are architected for deployment into regulated environments where data residency, access controls, and audit trails matter as much as accuracy metrics. This positioning addresses the gap between frontier model capabilities and enterprise operational requirements: latency SLAs, cost predictability, and compliance frameworks that prevent many organizations from operationalizing public AI APIs. Cohere Labs has published over 100 papers and built a research community of 4,500+ researchers, signaling ongoing investment in foundational work rather than pure application-layer focus. The team composition skews heavily toward researchers and engineers from academic backgrounds, which maps to the technical challenge space - building models that balance performance, safety constraints, and deployment flexibility across varied enterprise infrastructure.
Perplexity
Perplexity operates an AI-powered answer engine processing over 150 million questions weekly across web, mobile, and enterprise platforms. Founded in 2022, the system combines real-time web search with multiple LLMs to deliver source-attributed answers. The architecture serves both consumer and enterprise workloads, with enterprise deployments requiring security guarantees for knowledge worker use cases including legal research partnerships with organizations like Latham & Watkins. The technical stack runs on AWS infrastructure with Terraform for provisioning, Python and Go for backend services, and PyTorch with DeepSpeed and FSDP for model training and inference. Data pipelines use dbt, SQL, Snowflake, and Databricks. Frontend implementations use React and TypeScript, with Docker containerization and Open Policy Agent for access control. This architecture must handle tail latency and throughput requirements for real-time search retrieval paired with LLM inference at consumer scale, while maintaining source credibility verification in the critical path. The engineering focus centers on information retrieval accuracy, model response quality, and citation reliability rather than advertising optimization. Production systems must balance inference cost against answer quality across multiple models, manage retrieval latency for real-time web indexing, and maintain reliability for both free-tier consumer traffic and enterprise SLA requirements. Pro tier monetization suggests capacity-based or model selection tiering rather than pure ad-based revenue.
Descript
Descript builds a video and audio editing platform that replaces timeline-based manipulation with text-based editing - users cut and rearrange content by editing transcribed text rather than working directly with waveforms or video tracks. The system serves millions of creators, handling the full production pipeline from recording through collaborative editing to publication. Core technical domains span machine learning for transcription and automated design, text-based editing interfaces built on React and TypeScript, and distributed collaboration infrastructure. The platform's architecture supports both solo and team workflows across time zones, with backend systems running on PostgreSQL and Redis. Technical focus areas include generative AI capabilities that create content from natural language descriptions, automated design systems that reduce manual formatting work, and the fundamental text-to-media mapping that enables document-style editing of temporal content. The team combines creator domain expertise with systems engineering - reflected in stated priorities around human-centered design and products that handle real production constraints rather than demo cases. The stack centers on TypeScript/React for client interfaces, Python for ML pipelines, and SQL-based data infrastructure with dbt for transformation logic. REST APIs provide integration points. Current engineering emphasis appears weighted toward extending ML capabilities - transcription accuracy, generative features, design automation - alongside the operational complexity of maintaining reliable performance at scale for collaborative real-time editing workflows.