Responsibilities:
Play a significant role in deploying and maintaining critical data pipelines in production.
Lead strategic technological initiatives and long-term plans from initial exploration and POC to going live in a hectic production environment.
Design infrastructural data services and coordinate with the Architecture team, R&D teams, Data Scientists and Product Managers to build scalable data solutions.
Work in Agile process with Product Managers, and other TLs.
Requirements
- Sc. in Computer Science, or equivalent
- 3+ years of extensive experience with programming languages (preferably, Python) – a must!
- 3+ years of extensive SQL experience (preferably working in a production environment) – a must!
- Strong capability of schema design and data modeling
- Experience in building robust and scalable data pipelines in a micro-services environment
- Experience with data services orchestration tools, such as Airflow
- Quick, self-learning and good problem-solving capabilities
- Good communication skills and collaborative
- Process and detailed oriented
- Passion to solve complex data problems
Desired:
- Experience with Snowflake and MSSQL
- Experience with MLOps and ML implementations
- Experience with Docker and Kubernetes
- Experience with GCP services
- Experience with PubSub/Kafka