Share
Key job responsibilities
• Own the design, development, and maintenance of last mile data sets
• Manipulate/mine data from database tables (Redshift, Apache Spark SQL)
• Conduct deep dive investigations into issues related to incorrect and missing data
• Identify and adopt best practices in developing data pipelines and tables: data integrity, test design, build, validation, and documentation.
• Continually improve ongoing reporting and data processes in AMZL
• Work with in-house scientists, global supply chain, transportation and logistics teams, and software teams to identify new features and projects.
• Identify ways to automate complex processes through AWS.
• This is an individual contributor role that will partner with internal stakeholders across multiple teams, gathering requirements and delivering complete solutions
- 3+ years of data engineering experience
- Experience with data modeling, warehousing and building ETL pipelines
- Experience in at least one modern scripting or programming language, such as Python, Java, Scala, or NodeJS
- Experience with one or more query language (e.g., SQL, PL/SQL, DDL, MDX, HiveQL, SparkSQL, Scala)
- Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions
- Experience with big data technologies such as: Hadoop, Hive, Spark, EMR
- Speak, write, and read fluently in Japanese
These jobs might be a good fit