Act as a primary point of contact for production support of our business data systems and ETL pipelines.
Proactively monitor system performance, data loads, and pipeline health to identify and resolve issues before they impact our business users.
Investigate, troubleshoot, and perform root cause analysis for system failures, data discrepancies, and performance bottlenecks.
Develop and enhance operational tools, dashboards, and automated scripts (primarily in Python and SQL) to improve system monitoring and support efficiency.
Collaborate closely with Engineering Program Managers (EPMs), data engineers/SWEs, QA, and product managers to ensure system stability and resolve complex issues.
Document technical issues, solutions, and failure scenarios to build a robust knowledge base for the team.
Respond to inquiries from business users regarding data availability and system issues in a timely and professional manner.
4+ years of hands-on experience in a data-centric role with a strong emphasis on production support, data engineering, or a similar field.
BS or MS in Computer Science, Engineering, or a related field, or equivalent practical experience.
Excellent problem-solving and analytical skills, with the ability to work independently in a fast-paced, dynamic environment.
Strong verbal and written communication skills, capable of explaining technical issues to both technical and non-technical partners.
A collaborative, team-oriented mindset with a passion for ensuring high-quality data and system reliability.
Proficiency in advanced SQL, including performance tuning and writing complex queries.
Experience with scripting languages for automation and ETL development, particularly Python (Shell or Golang is a plus).
Solid understanding of relational databases (e.g., Postgres, Oracle, Snowflake) and data warehousing concepts.
Hands-on experience working with and maintaining ETL pipelines and workflow management tools (e.g., Airflow).
Familiarity with version control systems (Git) and CI/CD practices.
Experience with cloud platforms (AWS, Google Cloud) and working with APIs is beneficial.
Familiarity with Big Data technologies (e.g., Spark, Hadoop, Hive) is a plus.
Exposure to AI/ML tools for operational monitoring or data analysis is a plus.