Finding the best job has never been easier
Share
Key job responsibilities2. Actively participates in the code review process, design discussions, team planning, operational excellence, and constructively identifies problems and proposes solutions
3. Makes appropriate trade-offs, re-use where possible, and is judicious about introducing dependencies
4. Makes efficient use of resources (e.g., system hardware, data storage, query optimization, AWS infrastructure etc.)
5. Knows about recent advances in distributed systems (e.g., MapReduce, MPP Architectures, External Partitioning)
6. Asks correct questions when data model and requirements are not well defined and comes up with designs which are scalable, maintainable and efficient
7. Makes enhancements that improve team’s data architecture, making it better and easier to maintain (e.g., data auditing solutions, automating, ad-hoc or manual operation steps)
8. Owns the data quality of important datasets and any new changes/enhancementsA day in the life
This role requires you to live at the intersection of data, software, and analytics. We leverage a comprehensive suite of AWS technologies, with key tools including S3, Redshift, DynamoDB, Lambda, API's, Glue. You will drive the development process from design to release.
Managing data ingestion from heterogeneous data sources, with automated data quality checks.
Creating scalable data models for effective data processing, storage, retrieval, and archiving.
Using scripting for automation and tool development, which is scalable, reusable, and maintainable.
Providing infrastructure for self serve analytics and science use cases.
Using industry best practices in building CI/CD pipelines
- 3+ years of data engineering experience
- Experience with data modeling, warehousing and building ETL pipelines
- Experience with SQL
These jobs might be a good fit