Software Engineer, ML Infrastructure

Anysphere, Inc.
San Francisco, United States of America
1 month ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English

Job location

San Francisco, United States of America

Tech stack

Configuration Management
ETL
Cursor (Graphical User Interface Elements)
Linux
Distributed Data Store
InfiniBand
Python
Cloud Services
Software Engineering
Software Systems
TypeScript
Rust
Graphics Processing Unit (GPU)
Reliability of Systems
Kubernetes
Bare Metal
Slurm
Go

Job description

The ML Infrastructure team builds large-scale compute, storage, and software infrastructure to support Cursor's work building the world's best agentic coding model. We're looking for strong engineers who are interested in building high-performance infrastructure and the software to support it. This role works closely with ML researchers and engineers to enable their work through improvements to our training framework, systems reliability/performance, and developer experience., * Collaborate with ML researchers to improve the throughput and reliability of training

  • Work with OEMs, cloud service providers, and others to plan and build cutting-edge GPU infrastructure
  • Improve the density and scalability of compute environments to enable increasingly large RL workloads
  • Create software and systems to automate building, monitoring, and running GPU clusters
  • Build workload scheduling and data movement systems to support Cursor's growing training footprint

Requirements

Do you have experience in TypeScript?, * A strong background in systems and infrastructure-focused software engineering, particularly in Python, Typescript, Rust, and Golang

  • Experience with distributed storage and networking infrastructure, particularly on Linux systems across cloud and bare metal environments
  • Exposure to large-scale systems and their unique challenges, ideally across thousands of nodes with significant resource footprints.
  • Production use of infrastructure-as-code and configuration management, across hosts and Kubernetes, * Operational exposure to Nvidia GPUs with Infiniband or RoCE, particularly with Blackwell and Hopper-class hardware
  • Exposure to Ray, Slurm, or other common compute and runtime schedulers

Apply for this position