- Home
- AI Companies
- Harrison.ai
Harrison.ai
Similar companies
Graphcore
Graphcore, a British semiconductor company and wholly owned subsidiary of SoftBank Group, develops specialized AI compute hardware centered on its Intelligence Processing Unit (IPU). The IPU represents a processor architecture specifically designed for machine intelligence workloads rather than general-purpose computing. The company built a complete AI compute stack spanning silicon design through datacenter infrastructure, including the Poplar software framework that sits atop the hardware. Graphcore brought the first Wafer-on-Wafer AI processor to market, a packaging approach that addresses the bandwidth and latency constraints inherent in traditional chip-to-chip interconnects for AI workloads. The technical scope encompasses semiconductor engineering, processor design, and AI-specific optimizations across both hardware and software layers. The engineering team works on silicon design, wafer-scale integration technology, and the development of tools for AI model optimization. The software stack includes developer tools designed to extract performance from the IPU architecture, with ongoing work to optimize popular AI models for the platform. This systems-level approach attempts to address the throughput and efficiency bottlenecks that emerge when running large-scale machine learning workloads on conventional processor architectures. Under CEO Nigel Toon's leadership, Graphcore operates with global presence and maintains teams of semiconductor, software, and AI specialists. The company's technology stack includes standard datacenter interfaces (PCIe, DDR, Ethernet) alongside proprietary elements like the IPU and Poplar software. The subsidiary structure under SoftBank provides backing for continued development of both the silicon and the software layers required to compete in AI compute infrastructure, where the trade-offs between custom silicon development costs and performance gains define commercial viability.
Vast.ai
Vast.ai operates a peer-to-peer GPU marketplace connecting over 10,000 GPUs across 40 data centers with users requiring compute for training, fine-tuning, and inference workloads. The platform aggregates capacity from data centers and individual providers running Vast's hosting software, offering on-demand, interruptible, and auction-based pricing models that price 3-5x below traditional cloud providers. Instance deployment occurs in seconds, with the marketplace enabling direct comparison of price-performance across heterogeneous hardware. The architecture surfaces a pricing-availability trade-off inherent to peer-to-peer models: cost savings derive from utilizing underutilized capacity, but availability and reliability vary by provider. Interruptible instances present the sharpest cost-performance point but require fault-tolerant workloads and checkpointing discipline. The platform supports standard ML frameworks (PyTorch, TensorFlow) and containerized deployments via Docker. Enterprise offerings provide dedicated clusters with SLAs, SOC 2 Type I certification, and access to ISO 27001 certified facilities, trading marketplace economics for operational predictability. The technical stack spans Python and C++ for core platform services, PostgreSQL for marketplace state, Redis for coordination, and Terraform for infrastructure provisioning. CUDA support is foundational for GPU workloads. The system must handle heterogeneous provider configurations, node churn, and pricing dynamics across thousands of GPUs while maintaining search and allocation latency suitable for rapid instance provisioning. Founded in 2018, the company positions itself as infrastructure for cost-sensitive training and inference at scale.
Pinecone
Pinecone operates a fully managed vector database service designed for production AI applications requiring storage and retrieval of high-dimensional embeddings. The system handles vector search at scale across recommendation systems, semantic search, and related ML-backed services. Founded by Edo Liberty, formerly a research director at AWS with prior experience building custom vector search systems at large scale, the company is credited with establishing the vector database category as a distinct infrastructure layer. The technical stack centers on systems languages - Rust, Go, C++, and Python - with RocksDB as the storage engine and Kubernetes orchestration across AWS, GCP, and Azure. This architecture targets the operational complexity of managing embedding indices, query latency, and throughput at production scale, abstracting infrastructure decisions from engineering teams deploying AI features. The platform serves thousands of companies, positioning itself on ease of deployment and reduced time-to-production for vector-backed applications. The founding principle emphasizes accessibility for engineering teams of varying sizes, evolving the managed service model to minimize operational overhead in running vector workloads. Core focus areas include retrieval performance, reliability under production load, and cost-efficiency trade-offs inherent to high-dimensional search systems.
OpenEvidence
OpenEvidence operates a HIPAA-compliant medical information platform that handles over 100 million AI-powered clinical consultations from U.S. doctors and frontline clinicians. The system functions as a natural language search and retrieval layer over medical literature, synthesizing evidence from trusted sources to deliver point-of-care clinical decision support. With more than 40% of U.S. physicians logging in daily, the platform addresses a core bottleneck in clinical workflows: the exponential growth of medical knowledge against fixed physician time budgets. The system surfaces relevant evidence in seconds rather than the hours traditional literature review requires. The technical architecture supports evidence synthesis across landmark medical literature, aggregating content through clinical partnerships while maintaining compliance constraints required for healthcare settings. The platform serves as a knowledge management system that operates across practice environments - from academic medical centers to rural clinics - suggesting infrastructure designed for variable network conditions and diverse deployment contexts. Free access for verified U.S. healthcare professionals indicates a distribution model optimized for maximum clinician adoption rather than per-seat pricing common in enterprise healthcare software. Core technical domains span clinical decision support, AI copilot functionality for clinicians, and content aggregation from medical literature sources. The system's reliability requirements are elevated given its role in clinical decision pathways affecting patient outcomes, demanding careful evaluation of failure modes where incorrect or incomplete evidence synthesis could influence treatment decisions.