Key job responsibilities
* Design, implement, and automate deployment of our distributed system for collecting and processing log events from multiple sources
* Manage Petabyte scale Database for the organization
* Design data schema and operate internal data warehouses and SQL/NoSQL database systems
* Own the design, development, and maintenance of ongoing metrics, reports, analyses, and dashboards to drive key business decisions
* Monitor and troubleshoot operational or data issues in the data pipelines
* Drive architectural plans and implementation for future data storage, reporting, and analytic solutions
* Work collaboratively with business analysts, data scientists, and other internal partners to identify opportunities/problems
* Provide assistance to the team with troubleshooting, researching the root cause, and thoroughly resolving defects in the event of a problem
- 1+ years of data engineering experience
- 2+ years of analyzing and interpreting data with Redshift, Oracle, NoSQL etc. experience
- Bachelor's degree in a quantitative/technical field such as computer science, engineering, statistics
- Knowledge of distributed systems as it pertains to data storage and computing
- Experience with data modeling, warehousing and building ETL pipelines
- Experience with one or more query language (e.g., SQL, PL/SQL, DDL, MDX, HiveQL, SparkSQL, Scala)
- Experience with one or more scripting language (e.g., Python, KornShell)
- Experience with big data technologies such as: Hadoop, Hive, Spark, EMR
- Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc.
משרות נוספות שיכולות לעניין אותך