MLOps Engineer
Job Description
About this role
ML in production lives or dies by the unglamorous work of pipelines, registries, and rollback plans — and AI assistants tend to focus on the model and ignore everything around it. As an MLOps Engineer for AI training, you will help AI generate MLOps code that ships, scales, and recovers, not just the notebook code that demos well.
Key Responsibilities
• Generate and evaluate instruction-response pairs covering ML pipelines, feature stores, and model registries.
• Review AI-generated code using Kubeflow, MLflow, SageMaker, Vertex AI, and Azure ML.
• Provide feedback on CI/CD for ML, model packaging, and reproducibility.
• Validate AI handling of model serving (Triton, TorchServe, BentoML, KServe) and autoscaling.
• Evaluate observability for ML (Evidently, WhyLabs, custom monitoring) and drift detection.
• Identify subtle issues in pipeline DAGs, feature/serving skew, and shadow deployments.
Ideal Qualifications
• 5• years in MLOps, ML platform engineering, or production ML systems.
• Deep familiarity with Kubernetes-based ML serving and at least one major cloud ML platform.
• Strong grasp of Docker, Helm, and infrastructure-as-code for ML workloads.
• Experience with feature stores (Feast, Tecton) and model registries.
• Comfort with Python, plus a systems language (Go, Rust, or Java) for platform work.
• Familiarity with GPU scheduling and distributed training orchestration is a plus.
Project Timeline
• Start Date: Immediate
• Duration: Ongoing
• Commitment: Flexible, 10-25 hours/week
Contract & Payment Terms
• Independent contractor agreement
• Remote work — anywhere in eligible locations
• Weekly payment via Stripe or bank transfer
• Flexible hours
Ship AI with AI — apply now!