- Home
- AI Companies
- IREN
IREN
About
This company hasn't shared a description yet.
Similar companies
Nebius
Nebius is a Nasdaq-listed technology company (NBIS) building full-stack AI infrastructure from its Amsterdam headquarters, with GPU clusters deployed across Europe and the United States. Led by CEO Arkady Volozh, the company operates AI-optimized sustainable data centers - including a facility 60 kilometers from Helsinki and a new Vineland, New Jersey site - and has raised significant capital ($700 million from investors including Accel, NVIDIA, and Orbis). The engineering organization, numbering in the hundreds, maintains deep expertise in world-class infrastructure and runs an in-house AI R&D team that dogfoods the platform to validate it against production ML practitioner requirements. The infrastructure stack spans hyperscaler-scale features with supercomputer-grade performance characteristics. ISEG, Nebius's supercomputer, ranks among the world's most powerful systems. The platform integrates NVIDIA GPUs with NVIDIA InfiniBand networking, exposing workload orchestration through both Kubernetes and Slurm. The operational layer includes standard observability (Prometheus, Grafana), data infrastructure (PostgreSQL, Apache Spark), and ML tooling (MLflow, vLLM, Triton, Ray), with infrastructure-as-code managed via Terraform. This architecture targets the latency, throughput, and reliability requirements of AI training and inference workloads at scale. The company has secured a multi-billion dollar agreement with Microsoft to deliver dedicated AI infrastructure from its Vineland data center. Nebius serves startups, research institutes, and enterprises across healthcare and life sciences, robotics, finance, and entertainment verticals. The technical approach emphasizes production-grade infrastructure that handles the operational complexity of large-scale AI deployments - managing GPU utilization, network bottlenecks, and the cost-performance trade-offs inherent in serving diverse AI workloads from model training through inference serving.
Qdrant
Qdrant is a Rust-based vector database designed for high-dimensional similarity search at scale, serving semantic search, recommendation systems, and retrieval-augmented generation workloads. The system has processed billions of vectors across production deployments, with adoption reflected in 10 million+ downloads and 23,000 GitHub stars. The architecture trades language-level memory safety and zero-cost abstractions for predictable performance characteristics under load, operating both as an open-source deployment target and a managed cloud service. The database handles multi-modal retrieval and real-time recommendation workloads for enterprises including HubSpot, Bayer, Bosch, and CB Insights, spanning e-commerce through healthcare verticals. The managed offering positions deployment time as a primary bottleneck reducer, though actual production reliability depends on vector dimensionality, query patterns, and infrastructure topology. The team of 75+ distributed across 20+ countries maintains both the core engine and cloud operations, with the stack including gRPC for service boundaries, Kubernetes for orchestration, and observability through Prometheus/Grafana/OpenTelemetry. Founded in 2021 by André Zayarni and Andrey Vasnetsov, the company operates a dual open-source and managed cloud business model. The technical focus centers on scalability trade-offs in nearest neighbor search - balancing index structure overhead, query latency distribution, and write throughput as vector counts scale. Deployment options span AWS, GCP, and Azure, with Terraform for infrastructure provisioning and Docker for containerization.
Mirelo AI
Mirelo AI builds foundation models for generating synchronized audio for video content, targeting the latency and quality bottleneck in audio-for-video workflows. Founded in 2023 in Berlin, the company raised $41 million in seed funding co-led by Index Ventures and Andreessen Horowitz. Their models generate synchronized sound effects in seconds rather than the hours typically required for manual sound design, addressing production throughput constraints across gaming, film, social media, and broader visual content verticals. The technical stack centers on PyTorch with transformer architectures, optimized for H100 and H200 GPUs using Nsight profiling and SLURM for cluster orchestration. The team sources from Google Brain, Amazon, Meta FAIR, Disney, ETH Zürich, and Max Planck Institutes, combining AI research depth with domain expertise from musicians and product specialists. Co-founder and CEO CJ Simon-Gabriel previously worked at AWS Labs, where the founding team originated. The core technical challenge is tight audio-visual synchronization at generation time - a constraint that spans model architecture design, latency optimization, and evaluation methodology. Production systems must handle variable-length video inputs while maintaining temporal coherence across generated audio, requiring careful trade-offs between generation speed, output quality, and computational cost. The company positions its models as infrastructure for visual content pipelines, treating audio generation as a systems problem rather than a standalone creative tool.
Reka
Reka builds unified multimodal foundation models that process text, images, video, and audio. The company's core technical focus is modeling the physical world through systems that handle perception, reasoning, and action across modalities. The team includes researchers and engineers from Google DeepMind and Facebook AI Research working on inference-critical domains including GPU performance engineering, computer vision, audio processing, and natural language understanding. The technical stack centers on Python, PyTorch, and JAX for model development, with CUDA and C++ for performance-critical components. Infrastructure runs on Kubernetes and Slurm for orchestration and job scheduling. Engineering roles emphasize end-to-end ownership - individuals work across the stack from model architecture through deployment, addressing bottlenecks in latency, throughput, and operational complexity at production scale. Reka operates remote-first, aggregating global talent into a distributed systems organization. The work targets enterprise and organizational deployments where multimodal capabilities must meet reliability and cost constraints. Team structure reflects early-stage dynamics: engineers wear multiple hats, and technical decisions directly shape product capabilities and production characteristics.