Pipl is the identity trust company. Our solutions allow organizations to provide frictionless customer experiences and approve more transactions with greater confidence and speed. We use multivariate linking to establish deep connections among more than 330 billion trust signals—email, mobile phone and social media data that spans the globe—and then look at the big picture to derive identity trust.
We’re looking for a
Data Engineer to join our growing data team and help design, build, and scale our data infrastructure. Our team works closely with product, engineering, and data science to ensure reliable, high-quality data pipelines that power analytics, machine learning models, and data-driven decision-making.
As a Data Engineer, you will be responsible for creating and maintaining systems that collect, process, and store vast amounts of data, ensuring it is accessible, reliable, and optimized for performance across the organization.
Responsibilities
- Design, build, and maintain scalable ETL pipelines from multiple sources.
- Work closely with product managers, data scientists, and analysts to ensure data solutions meet business and technical needs.
- Ensure data integrity, accuracy, and security across platforms, implementing monitoring and validation frameworks.
- Improve data pipeline efficiency and performance, ensuring low latency and cost-effective solutions.
- Recommend and implement new technologies, tools, and best practices for data engineering.
Requirements:
Requirements
- 4+ years of experience as a Data Engineer (or in a similar role).
- Strong programming skills in Python.
- Hands-on experience with Cloud environment (GCP is an advantage).
- Experience working with dbt for data transformations and modeling.
- Solid experience with SQL and database design (both relational and NoSQL).
- Proven track record in building and maintaining large-scale data pipelines using frameworks such as Spark, Airflow, Kafka, or similar.
- Strong understanding of data modeling, warehousing, and ETL best practices.
- Self-motivated, detail-oriented, and able to work autonomously.
- Excellent communication skills in English (Hebrew is an advantage).
Advantage
- Experience with containerization and orchestration (Docker, Kubernetes).
- Programming skills in Java and Scala.
- Familiarity with real-time data processing systems.
- Exposure to data security and compliance best practices.
- Prior experience working in a big data or search engine environment.
- Hands-on experience with Aerospike.