DevJobs

Senior MLops Engineer

Overview
Skills
  • Python Python
  • PyTorch PyTorch
  • TensorFlow TensorFlow
  • DevOps DevOps ꞏ 4y
  • CI/CD CI/CD
  • AWS AWS
  • GCP GCP
  • Azure Azure
  • Docker Docker
  • Kubernetes Kubernetes
  • Terraform Terraform
  • Ansible Ansible
  • Infrastructure Engineering ꞏ 4y
  • MLOps ꞏ 4y
  • ML Engineering ꞏ 4y
  • LangFuse
  • Observability
  • Model Monitoring
  • Model Drift Detection
  • MLfLow
  • Machine Learning Models
  • LLM Providers
  • High-throughput Inference
  • GPU Scheduling
  • Distributed Training
  • ISO 27001
  • HuggingFace
  • Embeddings
  • Retrieval Pipelines
  • SOC2
  • CloudFormation
  • Vector Databases
Cyera is on a mission to protect one of the world's most valuable resources: data. Our AI-native platform gives organizations a complete view of where their data lives, how it’s used, and how to keep it safe, so they can reduce risk and unlock the full value of their data, wherever it is.

Since our founding in 2021, we’ve grown fast- Cyera-fast - securing over $1.3 billion in funding from the biggest pockets on the planet and establishing a global team. Today, Cyera is the fastest growing data security company on the planet, trusted by the Fortune 500 and beyond.

As we expand our AI-powered capabilities, we are looking for an experienced MLOps Engineer to build, scale, and operate the ML infrastructure that powers our data intelligence engine. This is a high-impact role that sits at the intersection of ML, infrastructure, and security—responsible for enabling reliable, scalable, and efficient model development, training, deployment, and monitoring.

You will work closely with our Data, AI, DevOps, and Backend teams to architect next-generation ML systems that support large-scale processing, model lifecycle automation, and cloud-native inference across AWS, GCP, and Azure.

Build & Operate ML Infrastructure

  • Design, implement, and maintain the infrastructure that enables reliable, scalable ML operations across the organization.
  • Maintain core ML platform services such as model registries, artifact stores, monitoring, and observability, enabling experiment tracking and model versioning systems.

Enable Scalable Inference & Automation

  • Architect CI/CD processes for ML models, including automated testing, benchmarking, validation, and deployment.
  • Integrate ML workflows into our microservices-based platform and multi-cloud architecture.
  • Optimize real-time inference performance, cost efficiency, and reliability.

Observability, Reliability & Governance

  • Build monitoring, alerting, and logging solutions for model performance, drift, data quality, and system health.
  • Work closely with security teams to ensure compliance, governance, and risk mitigation for ML assets.
  • Troubleshoot and resolve production issues across distributed systems with petabyte-scale data.

Cross-Team Collaboration

  • Partner with Data Scientists and ML Engineers to streamline research-to-production workflows.
  • Work with DevOps and Backend teams on resource planning, automation, and cloud infrastructure optimization.
  • Participate in architectural POCs and contribute to the evolution of our AI platform.

Requirements:

  • 4+ years of experience in MLOps, ML Engineering, DevOps, or infrastructure engineering supporting ML workloads.
  • Practical experience deploying machine learning models and LLM Providers in production environments at scale.
  • Experience in cloud-native environments (AWS, GCP, Azure) — including compute, storage, networking, and security configurations.
  • Hands-on experience with containerized environments and orchestration (Docker, Kubernetes).
  • Strong understanding of CI/CD concepts and experience building automated pipelines for ML systems.
  • Experience with IaC tools (Terraform, CloudFormation, or Ansible).
  • Familiarity with distributed training, GPU scheduling, and high-throughput inference.
  • Experience with model monitoring, observability, or model drift detection.
  • Experience implementing research & ML tooling for experimentation (MLfLow, LangFuse)
  • Excellent problem-solving skills and the ability to work in a fast-paced, high-scale environment.
  • Strong communication and collaboration skills, with the ability to work across R&D teams.

Advantage:

  • Experience with LLM fine-tuning, deployment, or optimization.
  • Background in building internal ML platforms or ML tooling.
  • Strong Python fundamentals and hands-on experience with ML frameworks (PyTorch, HuggingFace, TensorFlow).
  • Experience managing petabyte-scale data systems.
  • Knowledge of vector databases, embeddings, or retrieval pipelines.
  • Familiarity with SOC2 / ISO 27001 requirements related to ML workflows.
  • Experience in highly scalable SaaS products and/or security products.

Why Join Us?



At Cyera, we care about collaboration, innovation, and agility. We take “teamwork” seriously—with our inclusive and supportive culture at the forefront—and we’re just as serious about nurturing Cyerans to grow, both personally and professionally.

Feel free to apply even if your experience doesn’t tick every box.

We’re building something special here—and we welcome Cyerans with diverse backgrounds, perspectives, and experiences.
Cyera