As a Data Engineer at JPMorgan Chase within the International Consumer Bank, you will be a part of a flat-structure organization. Your responsibilities are to deliver end-to-end cutting-edge solutions in the form of cloud-native microservices architecture applications leveraging the latest technologies and the best industry practices. You are expected to be involved in the design and architecture of the solutions while also focusing on the entire SDLC lifecycle stages.
Job responsibilities:
- Work on AWS cloud,, S3, AWS glue catalog, and snowflake tech teams to publish the data right
- Create plans for the development and delivery of product data to support strategic business objectives, business operations, advanced analytics, and metrics and reporting.
- Work with key partners to drive an understanding of the data and its use within the business. Provide subject matter expertise with respect to the content and use of data in the product and associated business area.
- Identify the scope of critical data within their product (inflows/outflows and key dependencies), ensuring that the prioritized data is well-documented as to its meaning and purpose, and classified accordingly with metadata to enable its understanding and control
- Support the aligned Data & Analytics lead for their product by identifying data required to be integrated into analytics platforms to support analytics projects.
- Document requirements for the accuracy, completeness, and timeliness of data within the product, and coordinate resources to deliver data quality requirements. Develop processes and procedures to identify, monitor, and mitigate data risks for data in the product, including risks related to data protection, data retention and destruction, data storage, data use, and data quality
Required qualifications, capabilities and skills:
- Academic qualification in a computer science or STEM (science, technology, engineering or mathematics) related field or the foreign equivalent
- Professional experience working in an agile, dynamic and customer facing environment
- Recent hands-on professional experience (actively coding) working as a data engineer (back-end software engineer considered)
- Understanding of distributed systems and cloud technologies (AWS, GCP, Azure, etc.)
- Understanding of data streaming and scalable data processing frameworks (Kafka, Spark Structured Streaming, Flink, Beam etc.)
- Experience with SQL (any dialect) and Data tools (ie. Dbt)
- Experience in the all stages of software development lifecycle (requirements, design, architecture, development, testing, deployment, release and support)
- Experience with large scale datasets , data lake and data warehouse technologies on at least TB scale (ideally PB scale of datasets) with at least one of {BigQuery, Redshift, Snowflake}
- Experience in Infrastructure as Code (ideally Terraform) for Cloud based data infrastructure
- Good experience with using a JVM language (Java/Scala/Kotlin, preferably Java 8+) or extensive knowledge of Python
Preferred qualifications, capabilities and skills
- Experience with a scheduling system (Airflow, Azkaban, etc.)
- Understanding of (distributed and non-distributed) data structures, caching concepts, CAP theorem
- Understanding of security frameworks / standards and privacy
- Desired – experience in automating deployment, releases and testing in continuous integration, continuous delivery pipelines
- A solid approach to writing unit level tests using mocking frameworks, as well as automating component, integration and end-to-end tests
- Experience with containers and container-based deployment environment (Docker, Kubernetes, etc.)