About the Role
- - - - What the Candidate Will Do ----
- Responsible for defining the Source of Truth (SOT), Dataset designfor multiple Uber teams.
- Identify unified data models collaborating with Data Science teams
- Streamline data processing of the original event sources and consolidate them in source of truth event logs
- Build and maintain real-time/batch data pipelines that can consolidate and clean up usage analytics
- Build systems that monitor data losses from the different sources and improve the data quality
- Own the data quality and reliability of the Tier-1 & Tier-2 datasets including maitaining their SLAs, TTL and consumption
- Devise strategies to consolidate and compensate the data losses by correlating different sources
- Solve challenging data problems with cutting edge design and algorithms.
- - - - Basic Qualifications ----
- 3+ years of Data engineering experience
- Demonstrated experience of working with large data volumes and backend services.
- Good working knowledge of SQL (mandatory) and any other languages ( Java, Scala, Python)
- ̇Working Experience of ETL, Data pipelines, Data Lake, Data Modeling fundamentals.
- Good problem solving and analytical skills
- Good team player and collaboration skills.
- - - - Preferred Qualifications ----
- Experience in data engineering and working with Big data
- Experience with ETL or Streaming data and one or more of, Kafka, HDFS, Apache Spark , Apache Flink , Hadoop
- Good to have experience backend services and familiarity with one of the cloud platform ( AWS/ Azure / Google /Oracle cloud)
* Accommodations may be available based on religious and/or medical conditions, or as required by applicable law. To request an accommodation, please reach out to .