DevJobs

Data Architect

Overview
Skills
  • Spark Spark
  • PyTorch PyTorch
  • ML ML
  • RDBMS RDBMS ꞏ 6y
  • Microservices Microservices ꞏ 8y
  • OOP OOP
  • Design Patterns
  • AWS AWS ꞏ 6y
  • Snowflake Snowflake ꞏ 6y
  • GCP GCP ꞏ 6y
  • Kubernetes Kubernetes ꞏ 8y
  • Cloud Technologies ꞏ 8y
  • Serverless ꞏ 8y
  • Event-Driven Architectures ꞏ 8y
  • Vector Stores ꞏ 6y
  • Non-Relational Databases ꞏ 6y
  • Cloud-Based Data-Warehousing ꞏ 6y
  • BigQuery ꞏ 6y
  • Huggingface
  • Hadoop
  • Deep Learning Systems
  • Kubeflow
  • Data Systems
  • Code Infrastructure
  • Code Design
  • SageMaker
  • Scalable Systems
  • Software Engineering
  • Big Data Technologies
  • System Design
  • AI

What will you be doing?


You are a top-notch software engineer who has been architecting, designing, and building production-grade distributed software systems for years. You are passionate about working with cutting edge technologies in various aspects, from storage to AI to everything in between, in order to build large-scale systems and deliver business impact. You favor a hands-on approach to things and immersing yourself in the technical challenge ahead, while always having your eyes set on the business requirements and product needs, making sure your work reaches production.


What you'll be doing


As a Data Architect you will join the Data group in spearheading our AI-centric product comprising dozens of AI/ML/Data subsystems by architecting, designing, POC-ing, and providing technical leadership to team members. You’ll work closely with Product and Engineering teams on translating business requirements to technical solutions on the edge of AI.


What should you have?


  • Proven experience as a Tech Lead or Architect designing, analyzing, and troubleshooting large-scale production-grade data-intensive distributed systems
  • Deep understanding of modern backend paradigms such as Microservices, event-driven architectures, Serverless, and Kubernetes
  • 8+ years of experience with cloud technologies, preferably in AWS and GCP
  • 6+ years of experience working with relational and non-relational databases, vector stores, and petabyte-scale cloud-based data-warehousing solutions such as Snowflake or BigQuery
  • Extensive experience with machine learning systems and taking ML projects from inception to production, based on SageMaker, Kubeflow, or similar technologies
  • Strong knowledge in system design OOP, design patterns, and applying code infrastructure and code design in building scalable systems
  • Familiarity with Big Data technologies like Hadoop, Spark, etc.
  • Advantage – experience building production-grade deep learning systems using the Pytorch and Huggingface ecosystems

AUI