DevJobs

Big Data Team Leader

Overview
Skills
  • SQL SQL ꞏ 2y
  • Python Python ꞏ 2y
  • Kafka Kafka ꞏ 2y
  • Neo4j Neo4j
  • AWS AWS ꞏ 2y
  • Airflow Airflow ꞏ 2y
  • Terraform Terraform
  • MSK ꞏ 2y
  • VPC ꞏ 2y
  • S3 ꞏ 2y
  • Redshift ꞏ 2y
  • RDS ꞏ 2y
  • Presto ꞏ 2y
  • Kinesis ꞏ 2y
  • Glue ꞏ 2y
  • EMR ꞏ 2y
  • EC2 ꞏ 2y
  • Apache Spark ꞏ 2y
  • Pulumi
  • Cloudformation
  • AWS Neptun

Rise offers fully programmatic media solutions for publishers, empowering them to make informed business decisions through advanced data-powered solutions and AI models. Our platform seamlessly integrates with their content and enhances the user experience. With advancements perfectly tailored to meet advertising needs, publishers are enabled to maximize both revenues and profits.


We are looking for a Big Data Team Leader to join us and play a pivotal role in building and maintaining the foundational Data services, tooling, and automation for Rise's Ad Serving Platform.


In this dynamic and collaborative environment, you will work with cutting-edge technologies to manage a large volume of data, operate at a significant scale, and handle substantial traffic. Among your key responsibilities will be overseeing data pipelines and infrastructure, as well as automating and improving current workflows in both production and development environments.

Among your key responsibilities will be overseeing data pipelines and infrastructure, as well as automating and improving current workflows in both production and development environments.


Responsibilities:

  • Lead a team of both MLOps and Data engineer members.
  • Support and optimize existing processes with limited guidance.
  • Provision and maintain cloud providers' data infrastructure.
  • Monitor data production environments.
  • Assist group members in creating efficient queries.
  • Work collaboratively with other teams to develop creative solutions to problems.


Requirements​:

  • At least 2 years of experience as a Big Data Team Lead - MUST!
  • Strong knowledge and experience with Big Data tools like Apache Spark or Presto - MUST!
  • Experience at AWS cloud platform and its data lake formation and stack: S3, Airflow, Glue, EMR, Redshift, RDS, EC2, VPC
  • Coding proficiency in Python programming languages
  • Solid knowledge of Apache Airflow or alternative ETL tool
  • Production experience with managing data pipelines and streaming tools such as Kafka, MSK or Kinesis.
  • Deep understanding of structuring and querying data using Structured Query Language (SQL)


Advantages:

  • Hands-on experience with IAC tools like: Terraform, Cloudformation, Pulumi
  • Experience with vector databases like: neo4j, AWS Neptun etc.

Rise