About

Nebius is a Nasdaq-listed technology company (NBIS) building full-stack AI infrastructure from its Amsterdam headquarters, with GPU clusters deployed across Europe and the United States. Led by CEO Arkady Volozh, the company operates AI-optimized sustainable data centers - including a facility 60 kilometers from Helsinki and a new Vineland, New Jersey site - and has raised significant capital ($700 million from investors including Accel, NVIDIA, and Orbis). The engineering organization, numbering in the hundreds, maintains deep expertise in world-class infrastructure and runs an in-house AI R&D team that dogfoods the platform to validate it against production ML practitioner requirements.

The infrastructure stack spans hyperscaler-scale features with supercomputer-grade performance characteristics. ISEG, Nebius's supercomputer, ranks among the world's most powerful systems. The platform integrates NVIDIA GPUs with NVIDIA InfiniBand networking, exposing workload orchestration through both Kubernetes and Slurm. The operational layer includes standard observability (Prometheus, Grafana), data infrastructure (PostgreSQL, Apache Spark), and ML tooling (MLflow, vLLM, Triton, Ray), with infrastructure-as-code managed via Terraform. This architecture targets the latency, throughput, and reliability requirements of AI training and inference workloads at scale.

The company has secured a multi-billion dollar agreement with Microsoft to deliver dedicated AI infrastructure from its Vineland data center. Nebius serves startups, research institutes, and enterprises across healthcare and life sciences, robotics, finance, and entertainment verticals. The technical approach emphasizes production-grade infrastructure that handles the operational complexity of large-scale AI deployments - managing GPU utilization, network bottlenecks, and the cost-performance trade-offs inherent in serving diverse AI workloads from model training through inference serving.

Open roles at Nebius

Explore 316 open positions at Nebius and find your next opportunity.

NE

Help Desk Technician (Tier 1)

Nebius

Amsterdam, North Holland, Netherlands (On-site)

2mo ago
NE

Field Technical Lead – Data Center Deployments

Nebius

New Jersey, United States (On-site)

$120K – $170K Yearly2mo ago
NE

Detection & Response Manager

Nebius

Tel Aviv-Yafo, Tel Aviv District, Israel (On-site)

2mo ago
NE

HPC Specialist Solutions Architect

Nebius

United States + 1 more (Remote)

$225K – $315K Yearly2mo ago
NE

IT Data Center Technician

Nebius

Kansas City, Missouri, United States (On-site)

2mo ago
NE

Senior Technical Program Manager - New Data Center Launches

Nebius

Tel Aviv-Yafo, Tel Aviv District, Israel (On-site)

2mo ago
NE

Technical Product Manager - AI Cloud Network

Nebius

Amsterdam, North Holland, Netherlands (On-site)

2mo ago
NE

Senior Software Engineer (Backend)

Nebius

United States or Remote (United States)

$140K – $200K Yearly2mo ago
NE

Customer Engineer

Nebius

United States (Remote)

$225K – $275K Yearly2mo ago
NE

Field Marketing Manager - UKI

Nebius

United Kingdom + 1 more (Remote)

2mo ago
NE

Senior IT Infrastructure Engineer (Longcross)

Nebius

London, England, United Kingdom (On-site)

2mo ago
NE

IT Support Manager (Minneapolis, MN)

Nebius

Minnesota, United States (On-site)

2mo ago
NE

Senior Data Centre IT Technician (East London)

Nebius

London, England, United Kingdom (On-site)

2mo ago
NE

Mechanical Engineer - Data Centers

Nebius

Amsterdam, North Holland, Netherlands or Remote (Europe)

2mo ago
NE

Infrastructure Software Engineer

Nebius

United States (Remote)

$150K – $210K Yearly2mo ago
NE

Field Network Engineer

Nebius

Mäntsälä, Uusimaa, Finland (On-site)

2mo ago
NE

IT Support Manager (Tulsa, OK)

Nebius

Oklahoma, United States (On-site)

2mo ago
NE

Data Center IT Manager (Minneapolis, MN)

Nebius

Minneapolis, Minnesota, United States (On-site)

2mo ago

Similar companies

HA

HappyRobot

HappyRobot is an AI workforce platform founded in 2023 that builds autonomous agents to handle end-to-end operational work across phone, email, messaging, and documents. The company focuses on logistics and industrial operations - supply chains, freight, and businesses that move physical goods - where complex, patterned work spans multiple communication channels and document formats. Rather than augmenting human workflows, HappyRobot's system is designed to own complete tasks autonomously, operating as an AI-native OS for operations. The platform has been deployed across over 150 enterprise customers, including DHL and Ryder, and the company has raised $62 million from investors including Y Combinator and Andreessen Horowitz. The technical approach centers on building AI workers that can manage the operational complexity inherent in real-economy businesses: inbound calls that require looking up order status across internal systems, email threads with multi-party coordination, document processing that feeds into downstream workflows. The platform integrates natural language processing for conversational interfaces with document automation capabilities, handling the operational load that typically requires human judgment and context-switching. The stack is built on TypeScript, Next.js, and Go, suggesting a focus on both frontend orchestration and backend performance for production-scale operations. The founding team - Pablo Palafox, Javier Palafox, and Luis Paarup - brings backgrounds in engineering and logistics, positioning the company to understand both the technical constraints of building reliable AI systems and the operational bottlenecks in target industries. The company's positioning as AI-native reflects a systems-level bet: that automating operations requires rethinking the entire operational stack rather than bolting AI onto existing software workflows. For engineers, the work involves building agents that handle reliability and failure modes in production environments where downtime has direct business impact - missed shipments, delayed communications, operational backlogs.

95 jobs
DE

Decagon

Decagon builds a conversational AI platform designed to replace or augment legacy customer support systems by deploying intelligent AI agents across chat, email, and voice channels. The company positions its technology as infrastructure for delivering concierge-level customer experiences at scale, targeting brands looking to support, onboard, and retain customers without proportional headcount growth. Led by CEO Jesse Zhang and founded by serial entrepreneurs, Decagon operates from the US and focuses on addressing the operational constraints of traditional customer support systems. The platform's core technical approach centers on Agent Operating Procedures (AOPs), a natural-language-to-code compilation system that allows non-technical users to define agent behavior while preserving technical team control over guardrails, integrations, and versioning. This design addresses a common trade-off in AI tooling: enabling rapid iteration by domain experts without sacrificing reliability controls or introducing configuration drift. The agent orchestration layer spans multiple channels and claims to amplify CX team impact by 10x, though specific benchmarks around latency, accuracy, or failure rate are not publicly detailed. Decagon's technical domains span conversational AI, natural language processing, multichannel messaging infrastructure, and automation systems. The platform emphasizes runtime guardrails and version management as first-class concerns, reflecting a systems-oriented approach to production deployment. The company claims to deliver always-on, personalized service, positioning its agents as operational infrastructure rather than experimental tooling. For engineers evaluating opportunities, the technical challenges likely involve scaling context-rich, stateful interactions across channels while maintaining consistency, handling edge cases in natural language understanding, and building abstraction layers that balance expressiveness with safety.

89 jobs
QD

Qdrant

Qdrant is a Rust-based vector database designed for high-dimensional similarity search at scale, serving semantic search, recommendation systems, and retrieval-augmented generation workloads. The system has processed billions of vectors across production deployments, with adoption reflected in 10 million+ downloads and 23,000 GitHub stars. The architecture trades language-level memory safety and zero-cost abstractions for predictable performance characteristics under load, operating both as an open-source deployment target and a managed cloud service. The database handles multi-modal retrieval and real-time recommendation workloads for enterprises including HubSpot, Bayer, Bosch, and CB Insights, spanning e-commerce through healthcare verticals. The managed offering positions deployment time as a primary bottleneck reducer, though actual production reliability depends on vector dimensionality, query patterns, and infrastructure topology. The team of 75+ distributed across 20+ countries maintains both the core engine and cloud operations, with the stack including gRPC for service boundaries, Kubernetes for orchestration, and observability through Prometheus/Grafana/OpenTelemetry. Founded in 2021 by André Zayarni and Andrey Vasnetsov, the company operates a dual open-source and managed cloud business model. The technical focus centers on scalability trade-offs in nearest neighbor search - balancing index structure overhead, query latency distribution, and write throughput as vector counts scale. Deployment options span AWS, GCP, and Azure, with Terraform for infrastructure provisioning and Docker for containerization.

27 jobs
RE

Reka

Reka builds unified multimodal foundation models that process text, images, video, and audio. The company's core technical focus is modeling the physical world through systems that handle perception, reasoning, and action across modalities. The team includes researchers and engineers from Google DeepMind and Facebook AI Research working on inference-critical domains including GPU performance engineering, computer vision, audio processing, and natural language understanding. The technical stack centers on Python, PyTorch, and JAX for model development, with CUDA and C++ for performance-critical components. Infrastructure runs on Kubernetes and Slurm for orchestration and job scheduling. Engineering roles emphasize end-to-end ownership - individuals work across the stack from model architecture through deployment, addressing bottlenecks in latency, throughput, and operational complexity at production scale. Reka operates remote-first, aggregating global talent into a distributed systems organization. The work targets enterprise and organizational deployments where multimodal capabilities must meet reliability and cost constraints. Team structure reflects early-stage dynamics: engineers wear multiple hats, and technical decisions directly shape product capabilities and production characteristics.

3 jobs
WA

Wabi

Wabi is the first personal software platform, transforming how people interact with technology through AI-powered mini apps. With $20 million in pre-seed funding, the company has quickly established itself as a pioneer in the User-Generated Software (UGS) movement, enabling anyone to create, share, and remix personalized applications without writing code. Founded by Eugenia Kuyda, former CEO of Replika, Wabi is building what investors call the "YouTube of apps" - a social platform where millions of creators can build and distribute software tailored to individual needs, tastes, and contexts. The platform represents a fundamental shift from one-size-fits-all applications to truly personal software experiences. Rather than searching for apps that approximately match their needs, users describe their exact requirements in natural language, and Wabi generates custom mini apps optimized for their specific routines, preferences, and life situations. Operating with a lean team of 2-10 employees, Wabi is positioned at the forefront of AI-driven creativity, turning every user into a potential software developer and ushering in a new era where software is made for all of us, by all of us.