MI

About

This company hasn't shared a description yet.

Similar companies

RA

Reflection AI

Reflection AI develops open foundation models targeting superintelligent autonomous systems, with current work focused on autonomous coding as a path to broader cognitive automation. The company combines reinforcement learning and large language models to build systems capable of handling most cognitive work on a computer, positioning autonomous code generation as the bottleneck to unlock that capability. The team includes contributors to AlphaGo, AlphaZero, PaLM, GPT-4, and Gemini, bringing production experience across game-playing RL systems and frontier language models. This background suggests familiarity with the trade-offs in training large-scale models - compute efficiency, sample complexity, and the operational challenges of running RL at scale alongside supervised pretraining. Reflection's stated objective centers on keeping superintelligence open and accessible through open foundation models. For inference practitioners, this implies potential work on model architectures, training infrastructure, and deployment systems designed for broad distribution rather than proprietary deployment. The autonomous coding focus suggests evaluation infrastructure for code generation, likely including metrics beyond pass@k - compilation rates, execution correctness, and performance characteristics of generated code under real-world constraints.

49 jobs
SU

Suno

Suno operates a generative AI platform for text-to-music synthesis, producing complete songs with vocals and instrumentation from natural language prompts. The technical work centers on training models to understand melody - a non-trivial challenge requiring cross-domain expertise in audio engineering, machine learning, and music production. The company maintains three office locations across Cambridge, New York, and Los Angeles, staffing teams that combine musicians, audio engineers, and ML practitioners. The core infrastructure involves model training pipelines purpose-built for musical understanding rather than general audio synthesis. This requires handling the specific latencies and quality trade-offs inherent to generative audio: output coherence across time, harmonic consistency, and production values that meet consumer expectations for listenable music. The team's composition - technical staff with musical backgrounds - addresses the evaluation problem directly: musical quality metrics remain difficult to automate, making human judgment from domain experts operationally necessary. Suno's stated technical approach blends musical intuition with systems work, suggesting decision-making that weighs subjective quality alongside standard ML metrics. This creates tension between data-driven optimization and aesthetic judgment - a bottleneck common in creative ML applications where human preference doesn't reduce cleanly to loss functions. The platform targets democratized music creation, which implies scale requirements and cost constraints typical of consumer-facing generative AI: balancing inference costs against output quality while maintaining acceptable latency for interactive use.

36 jobs
SA

SambaNova

SambaNova builds a full-stack AI inference platform centered on custom dataflow chips (RDUs) and a three-tier memory architecture designed to address latency and energy efficiency bottlenecks in generative AI deployment. The architecture targets enterprise and government workloads requiring on-premises or sovereign deployment - fine-tuning open-source models behind customer firewalls with full data and model ownership retention. The platform powers sovereign AI data centers across Australia, Europe, and the UK, focusing on avoiding vendor lock-in to proprietary inference services. The technical approach uses custom dataflow technology rather than GPU-based architectures, trading off ecosystem maturity for claimed improvements in inference throughput and energy consumption at scale. The three-tier memory design addresses memory bandwidth constraints common in transformer inference. The platform supports PyTorch-based model fine-tuning and deployment workflows, with integration points through Python and C++ APIs. Operational complexity centers on full-stack ownership - hardware, software, and deployment infrastructure - requiring coordination across chip design, systems software, and model serving layers. The stack includes standard ML tooling (PyTorch, Python) alongside proprietary components for the RDU runtime and memory management. Build and CI infrastructure uses Bazel and CircleCI; artifact management through Google Artifact Registry and JFrog. The deployment model targets enterprises prioritizing data sovereignty over cloud-based inference APIs, introducing trade-offs in operational overhead versus control and latency predictability for on-premises workloads.

7 jobs