

Share
Your key responsibilities
Develop and maintain scalable ETL (Extract, Transform, Load) processes to efficiently extract data from diverse sources, transform it as required and load it into data warehouses or analytical systems.
Design and optimize database architectures and data pipelines to ensure high performance, availability and security while supporting structured and unstructured data.
Integrate data from multiple sources, including APIs, third-party services and on-prem/cloud databases to create unified and consistent datasets for analytics and reporting.
Collaborate with data scientists, analysts and business stakeholders to understand their data needs, develop data solutions and enable self-service analytics.
Develop automated workflows and data processing scripts using Python, Spark, SQL, or other relevant technologies to streamline data ingestion and transformation.
Optimize data storage and retrieval strategies in cloud-based data warehouses such as AWS Redshift, Google Big Query, or Azure Synapse, ensuring scalability and cost-efficiency.
Maintain and improve data quality by implementing validation frameworks, anomaly detection mechanisms and data cleansing processes.
Thoroughly tests code to ensure accuracy and alignment with its intended purpose. Reviews the final product with end users to confirm clarity and understanding, providing data analysis guidance as needed.
Offers tool and data support to business users and team members, ensuring seamless functionality and accessibility.
Conducts regression testing for new software releases, identifying issues and collaborating with vendors to resolve them and successfully deploy the software into production.
Skills and attributes for success
A team player with strong analytical, communication and interpersonal skills
Constantly updating yourself about new technologies in the market
A winning personality and the ability to become a trusted advisor to the stakeholders
To qualify for the role, you must have
Minimum 3 years of relevant work experience, with at least 1 year in designing and maintaining data pipelines, ETL processes and database architectures.
Bachelor’s degree (B.E./B.Tech) in Computer Science or IT, or Diploma in Data Science, Statistics, or related field.
Strong proficiency in SQL, Python, or Scala for data manipulation, automation and pipeline development.
Experience working with big data processing frameworks such as Apache Spark, Hadoop, or Kafka.
Hands-on experience with cloud-based data platforms such as AWS (Redshift, Glue), Google Cloud (Big Query, Dataflow), or Azure (Synapse, Data Factory).
Ideally, you’ll also have
Strong verbal and written communication, facilitation, relationship-building, presentation and negotiation skills.
Be highly flexible, adaptable, and creative.
Comfortable interacting with senior executives (within the firm and at the client)
What we look for
Strong teamwork, work ethic, product mindset, client centricity and a relentless commitment to EY values.
We offer a competitive remuneration package where you’ll be rewarded for your individual and team performance. Our comprehensive Total Rewards package includes support for flexible working and career development, and with FlexEY you can select benefits that suit your needs, covering holidays, health and well-being, insurance, savings and a wide range of discounts, offers and promotions. Plus, we offer:
Support, coaching and feedback from some of the most engaging colleagues around
Opportunities to develop new skills and progress your career
The freedom and flexibility to handle your role in a way that’s right for you
If you can confidently demonstrate that you meet the criteria above, please contact us as soon as possible.
Make your mark.
Apply now.
These jobs might be a good fit