MLOps Engineer managing AI pipelines for computer vision models. Involves end-to-end model lifecycle streamlining in a hybrid work environment.
Responsibilities
Own the end-to-end ML pipeline for computer vision: data prep, training, evaluation, model packaging, artifact/version management, deployment, and monitoring (local GPU cluster + GCP).
Design and maintain containerized workflows for multi-GPU training and distributed workloads (e.g., PyTorch DDP, Ray, or similar).
Build and operate orchestration (e.g., Airflow/Argo/Kubeflow/Ray Jobs) for scheduled and on-demand pipelines across on-prem and cloud.
Implement and tune resource allocation strategies based on current and upcoming task queues (GPU/CPU/memory-aware scheduling; preemption/priority; autoscaling).
Introduce and integrate monitoring/telemetry for:
job health and failure analysis (retry, backoff, alerts),
data/feature drift and model performance (precision/recall, latency, throughput),
infra metrics (GPU utilization, memory, I/O, cost).
Harden GCP environments (permissions, networks, registries, storage) and optimize for reliability, performance, and cost (spot/managed instance groups, autoscaling).
Establish model governance: experiment tracking, model registry, promotion gates, rollbacks, and audit trails.
Standardize CI/CD for ML (data/feature pipelines, model builds, tests, and canary/blue-green rollouts).
Collaborate with CV researchers/engineers to productionize new models and improve training throughput & inference SLAs.
Continuously improve documentation: update existing pipeline docs and produce concise runbooks, diagrams, and “how-to” guides.
Requirements
Hands-on MLOps experience building and running ML pipelines at scale (preferably computer vision) across on-prem GPUs and a public cloud (GCP preferred).
Strong with Docker and Docker Compose in local and cloud environments; solid understanding of image build optimization and artifact caching.
Proficiency with Python and Bash for pipeline tooling, glue code, and automation; Terraform for infra-as-code (GCP resources, IAM, networking, storage).
Experience with orchestration: one or more of Airflow, Argo Workflows, Kubeflow, Ray, or Prefect.
Senior Software Developer working on ML Infrastructure and Deployment at Verafin. Helping develop cutting - edge fraud detection tools alongside analytics teams using AWS and Terraform.
Machine Learning Engineer developing advanced SLAM systems for autonomous trucking environments at Bot Auto. Collaborating with cross - functional teams to optimize mapping solutions and ensure operational stability.
Graduate Deep Learning Algorithm Developer developing perception technologies for autonomous driving. Tackling challenges in object detection and 3D perception using state - of - the - art deep learning models.
Principal AI/ML Engineer leading the AI/ML infrastructure development for WEX's risk service needs. Focused on innovative engineering and technology solutions within a high - stakes environment.
AI/ML Engineer developing solutions in artificial intelligence for HPE. Responsible for conducting research, designing AI solutions, and mentoring team members.
Machine Learning Engineer focusing on modeling cancer cells and developing related tools. Collaborating with researchers and scientists to advance cancer treatment through ML.
Machine Learning Engineer II developing production - grade ML models for fraud detection at GEICO. Collaborating on system architecture and ensuring optimal performance of fraud assessment systems.
AI/ML Engineer III designing and architecting AI solutions at Hewlett Packard Enterprise. Collaborating with teams to drive innovation and tackle complex problems.
AI/ML Engineer deploying state - of - the - art AI models to solve real - world problems at Brain Co. Working in healthcare, government, and energy sectors for impactful results.
Trainer at WeAndTheMany facilitating learning by sharing experiences and creating interactive sessions. Engaging with students to enhance their skills and knowledge through dynamic teaching methods.