DevJobs

Senior Data Infrastructure Engineer

Overview
Skills
  • Bash Bash
  • Python Python
  • Spark Spark
  • PostgreSQL PostgreSQL
  • AWS AWS
  • GCP GCP
  • Kubernetes Kubernetes
  • Docker Docker
  • Argo Workflows
  • Delta Lake
  • Databricks
About Us


At Milestone, we’re building a robust, scalable, and intelligent data platform to power advanced analytics and machine learning across our business. Our infrastructure supports diverse workflows, from real-time decision-making to large-scale data transformations, all running across AWS and GCP environments. We’re looking for a Senior Data Infrastructure Engineer to join our growing team and help take our platform to the next level.



What You’ll Do


  • Design, build, and maintain a scalable data infrastructure across AWS and GCP.
  • Lead the development and orchestration of robust data pipelines using Argo Workflows and Dockerized components.
  • Build and optimize data transformations using Databricks and PySpark.
  • Manage and monitor data pipelines that write to PostgreSQL databases for downstream applications.
  • Collaborate with DevOps and software engineers to ensure data is accurate, timely, and accessible.
  • Drive infrastructure decisions with a focus on reliability, observability, and maintainability.
  • Participate in code reviews, set best practices, and mentor junior engineers.
  • Evaluate and integrate new technologies that enhance our data platform’s capabilities.



Tech Stack You’ll Work With


  • Workflow Orchestration: Argo Workflows, K8s, Docker
  • Big Data Processing: Databricks, PySpark
  • Programming Languages: Python (primary), Bash
  • Storage & Databases: PostgreSQL, S3, GCS, Delta Lake
  • Cloud Platforms: AWS, GCP


What We’re Looking For


  • 6+ years of experience in data engineering or data infrastructure roles.
  • Strong proficiency in Python and Spark (PySpark) for distributed data processing.
  • Strong knowledge of working with Databricks for data processing, analytics and ML.
  • Deep understanding of modern data architectures, including batch/streaming pipelines and lakehouse paradigms.
  • Hands-on experience managing cloud-native infrastructure on AWS and/or GCP.
  • Proficiency with PostgreSQL and understanding of how to optimize I/O from big data pipelines to relational databases.
  • Excellent communication skills and the ability to work cross-functionally.
  • Experience with Argo WorkflowsDocker, and orchestration of distributed data jobs (Optional).



Bonus Points


  • High proficiency with Delta Lake or Kubernetes.
  • Familiarity with CI/CD pipelines and infrastructure as code.
  • Previous experience in a startup or fast-paced product environment.



Why Join Us


  • Opportunity to shape core infrastructure powering real-world analytics products.
  • Ownership to make architecture and tooling decisions.
  • Collaborate with a sharp, mission-driven team that values engineering excellence.
  • Flexible work environment and competitive compensation package.
Milestone AI