Amp
About
This company hasn't shared a description yet.
Similar companies
Modal
Modal operates a serverless compute platform designed to minimize infrastructure friction for ML inference, fine-tuning, and batch workloads. The platform provides instant GPU access with usage-based pricing, targeting teams that need to ship compute-intensive applications without managing scheduling, container orchestration, or resource allocation. The architecture is built on custom infrastructure components - an in-house file system, container runtime, scheduler, and image builder - optimized for the latency and throughput characteristics of AI workloads. The technical stack spans Python, Rust, and Go at the systems level, with PyTorch, CUDA, vLLM, and TensorRT support for ML frameworks. This reflects prioritization of both developer ergonomics (Python interface) and low-level performance (Rust/Go for runtime components). The custom infrastructure signals investment in controlling the full vertical - from container initialization through GPU scheduling - rather than composing existing orchestration layers. The team operates across New York, Stockholm, and San Francisco, and includes creators of open-source projects like Seaborn and Luigi, alongside academic researchers and engineers with experience building production systems. The platform positions itself around developer experience as a core constraint, with infrastructure complexity abstracted to reduce operational overhead for data and AI teams.
Qdrant
Qdrant is a Rust-based vector database designed for high-dimensional similarity search at scale, serving semantic search, recommendation systems, and retrieval-augmented generation workloads. The system has processed billions of vectors across production deployments, with adoption reflected in 10 million+ downloads and 23,000 GitHub stars. The architecture trades language-level memory safety and zero-cost abstractions for predictable performance characteristics under load, operating both as an open-source deployment target and a managed cloud service. The database handles multi-modal retrieval and real-time recommendation workloads for enterprises including HubSpot, Bayer, Bosch, and CB Insights, spanning e-commerce through healthcare verticals. The managed offering positions deployment time as a primary bottleneck reducer, though actual production reliability depends on vector dimensionality, query patterns, and infrastructure topology. The team of 75+ distributed across 20+ countries maintains both the core engine and cloud operations, with the stack including gRPC for service boundaries, Kubernetes for orchestration, and observability through Prometheus/Grafana/OpenTelemetry. Founded in 2021 by André Zayarni and Andrey Vasnetsov, the company operates a dual open-source and managed cloud business model. The technical focus centers on scalability trade-offs in nearest neighbor search - balancing index structure overhead, query latency distribution, and write throughput as vector counts scale. Deployment options span AWS, GCP, and Azure, with Terraform for infrastructure provisioning and Docker for containerization.
Aleph Alpha
Aleph Alpha builds large language models and AI infrastructure designed for European governments and enterprises operating under strict sovereignty, compliance, and transparency requirements. Founded in 2019 and based in Germany, the company delivers PhariaAI, an end-to-end sovereign AI suite targeting public sector, industrial, and financial applications where data residency, intellectual property control, and regulatory alignment - particularly with the EU AI Act - are non-negotiable constraints. The technical focus centers on explainability primitives for LLMs and tokenizer-free architectures engineered to handle low-resource languages, addressing bottlenecks in both interpretability and multilingual coverage that standard transformer approaches struggle with in regulated environments. The architecture prioritizes auditability and operational control over raw performance. Aleph Alpha positions itself as an alternative to U.S.-centric AI ecosystems, explicitly targeting organizations seeking to avoid vendor lock-in while maintaining compliance with European regulations. The stack runs on Python, PyTorch, and Kubernetes, with deployment patterns optimized for on-premises and sovereign cloud environments where latency to external APIs or data egress to third-party infrastructure introduces unacceptable compliance risk. The company's technical domains span LLM research, explainability tooling, and platform development, with operational priorities weighted toward transparency, controllability, and alignment with organizational governance frameworks rather than maximizing throughput or minimizing inference cost in commodity cloud settings.