Data & ML Ops Engineer
Role details
Job location
Tech stack
Job description
As our Data & ML Ops Engineer, you will be driving the requirements gathering, development, rollout and operation of the related infrastructure. The systems you build and operate power every ML experiment, training run, and production deployment at Gravis. You will work at the intersection of our Platform, Autonomy, and Perception teams to enable high velocity & quality ML development & deployment., * Design and operate end-to-end ML pipelines covering data ingestion, preprocessing, versioning, training, evaluation, and deployment ranging from edge devices in the field to cloud training infrastructure
- Build and maintain a scalable data platform for large-scale multimodal robotics datasets (LiDAR point clouds, camera imagery, GNSS/IMU, and other machine data)
- Own CI/CD pipelines for ML workflows, including automated model training, regression testing, and staged deployment to production fleets
- Manage experiment tracking, model registry, and artifact versioning to ensure full reproducibility across research and production
- Collaborate closely with Autonomy and Perception engineers to understand requirements and translate them into reliable, scalable training environments
- Evaluate and integrate best-in-class MLOps tooling on cloud and on-prem compute platforms, We are an international team that is working to solve problems with a global impact: to facilitate efficient communication and collaboration, proficiency in English is a requirement for all roles.
Requirements
- Bachelor's or Master's degree in Computer Science, Data Engineering, Electrical Engineering, or a related field
- 3+ years of hands-on experience in ML Ops, data engineering, or ML infrastructure roles
- Strong Python skills and solid experience with ML frameworks such as PyTorch or TensorFlow
- Proven experience building and managing CI/CD pipelines for ML workloads (e.g. GitHub Actions or GitLab CI)
- Hands-on experience with containerization (Docker).
- Experience with cloud platforms (AWS, GCP, or Azure).
- Experience with data versioning, experiment tracking and workload orchestration tools (e.g. MLflow, W&B, clear.ml, DVC)., * Experience with GPU accelerated simulation environments (e.g. IsaacSim/IsaacLab, CARLA, MuJoCo)
- Experience working with robotics data (point clouds, camera streams, timeseries data).
- Hands-on experience with infrastructure as code
- Experience with Robotics & DevOps related tooling (Foxglove, Prometheus, Grafana)
- Experience scaling ML infrastructure
Benefits & conditions
This is an opportunity to join a dynamic and versatile team, and to be part of a young startup that will revolutionize heavy construction. Gravis Robotics offers a fair market salary and a working location in the vibrant city of Zurich. As a forward-facing startup, we understand that work-life balance and flexibility are important considerations for many professionals: If you are a highly qualified candidate with the requisite skills and experience, we encourage you to apply and discuss your preferred working arrangement during the interview process.