What You’ll Do- Design, develop, and maintain scalable data pipelines and backend services for processing large-scale web event analytics.
- Lead the development of core data infrastructure, ensuring high availability, performance, and reliability.
- Solve complex challenges at the intersection of low latency, high correctness, and full determinism.
- Architect and optimize data models, ETL/ELT workflows, and distributed data processing frameworks.
- Participate in code reviews, mentor junior engineers, and enforce best practices in data engineering.
- Diagnose and resolve data quality issues, ensuring accuracy, efficiency, and security.
What You’ll Need- 3-5 years of experience in data engineering or backend software development.
- Proficiency in Python, SQL, Java, or Scala for data pipeline development.
- Expertise in Big Data frameworks (e.g., Spark, Flink, Presto, MapReduce) and distributed computing.
- Strong SQL skills with experience in data modeling and query optimization.
- Hands-on experience with ETL pipeline development and data warehousing.
- Knowledge of Hadoop ecosystem technologies (HDFS, Hive, Oozie, Airflow).
- Ability to design and implement scalable data architectures that drive business impact.
Preferred Qualifications- Experience with real-time data streaming (e.g., Flink, Kafka).
- Background in AdTech or large-scale data analytics.
* Accommodations may be available based on religious and/or medical conditions, or as required by applicable law. To request an accommodation, please reach out to .