Representative role.This is a composite of searches we run regularly at this level and stack. If your background matches, apply anyway — we’ll match you to similar live briefs.
An ASX-listed company is building out the ML platform team and needs experienced MLOps engineers. The brief: make it easy for 20+ data scientists to ship models to production reliably.
The stack is AWS + Databricks + MLflow + Terraform + Kubernetes. The team is small but well-supported. The work is a mix of platform engineering (deploy pipelines, feature stores, monitoring) and customer success (unblocking data scientists who can't figure out why their model is slow).
What you’ll do
- Own the model deployment pipeline: training → registry → staging → production → monitoring
- Build and maintain infrastructure-as-code for ML workloads (Terraform, SageMaker, Databricks)
- Set up feature store, experiment tracking, and model monitoring (drift, latency, cost)
- Partner directly with data scientists — pair on deployments, debug production issues, teach patterns
- Own CI/CD for ML — from notebook to automated pipeline
What you bring
- 5+ years of DevOps/SRE/Platform Engineering, at least 2 of which on ML-specific workloads
- Deep AWS (SageMaker, ECS, Lambda, Step Functions) or GCP (Vertex AI) experience
- Terraform or equivalent IaC
- Python proficient — not just scripting; you can read and contribute to data-science codebases
- Kubernetes literacy
Nice-to-haves
- Databricks or Snowflake deep experience
- MLflow, Weights & Biases, or Kubeflow at scale
- Experience with model monitoring tools (WhyLabs, Arize, Fiddler)
