About the Role
- - - - Basic Qualifications ----
- Building optimized ETL Data Pipelines that are scalable to petabytes
- Using Strong Data Warehousing and data modeling concepts for easy consumption of data
- Strong SQL Skills
- Performance Troubleshooting and Tuning and ensuring the highest Data quality
- Should be self-motivated and passionate about bringing efficiency into the system through optimizations.
- Exposure to Hive, Spark, and data infra concepts
- BA or above in Comp Science or related field
- - - - Preferred Qualifications ----
- Experience with large-scale data warehousing architecture and data modeling
- Any Programming and scripting language (e.g. Python, GO, Java/Scala)
- Designing end-to-end data solutions and architecture with Data governance as the backbone
- Knowledge of Hadoop-related technologies such as HDFS, Hive/Spark, PySpark Presto, Hive, Airflow, Hudi, DeltaLake
- Data warehouses like Snowflake, Redshift, Big Query etc
- Realtime processing using Flink, Spark Streaming, Pinot, STORM
- Cloud technologies (GCP is a big plus)
For San Francisco, CA-based roles: The base salary range for this role is USD$158,000 per year - USD$175,500 per year.
You will be eligible to participate in Uber's bonus program, and may be offered an equity award & other types of comp. You will also be eligible for various benefits. More details can be found at the following link .