Job Description:
Software Development
- Write and maintain clean, efficient, and scalable code primarily in Java and Scala for data pipelines and microservices.
- Implement microservices architecture ensuring modularity, scalability, and maintainability.
- Conduct rigorous code reviews, debugging, and develop automated tests to verify software functionality and performance.
- Develop and maintain RESTful APIs, Java services, and batch processing jobs using Spring Framework(including Spring MVC, Spring Boot/Batch where applicable) to support scalable enterprise applications.
Data Pipeline and Processing
- Build and optimize ETL pipelines to process large-scale merchant data efficiently and reliably.
- Implement complex data transformations using Apache Spark and Google Cloud Dataproc.
- Develop and manage automated workflows using Google Cloud Dataflow and Apache Beam for seamless, scalable data orchestration and processing.
Cloud Infrastructure and Data Storage
- Build and optimize data storage solutions on Google BigQuery and BigTable for fast, reliable data querying.
- Monitor and maintain Google Cloud infrastructure focusing on security, availability, and performance.
Collaboration and Integration
- Participate actively in agile development processes including code reviews, daily stand-ups, and design discussions.
- Adhere to best practices in software engineering such as version control and continuousintegration/deployment.
Preferred Qualification:
Technical Skills
- Proficient in Java and/or Scala programming languages with solid software development principles.
- Familiarity of core computer science concepts such as data structures, algorithms, software design principles, object-oriented programming, and relational database concepts.
- Knowledge of ETL processes and data pipeline development.
- Knowledge of Google Cloud technologies including BigQuery, BigTable, DataProc, and Dataflow.
- Knowledge ofbuilding and optimizing data storage solutions and managing cloud infrastructure for high availability and performance.
- Knowledge of distributed data processing frameworks such as Apache Spark, Apache Beam.
- Proficient in Java and the Spring ecosystem with experience in building REST APIs, web services, scalable microservices and batch workflows.
- Familiarity with Python programming for data processing and automation to complement Java-based Spring applications.
- Familiarity with developing automated tests to ensure software functionality and performance.
- Ability to design, develop, and manage automated workflows for seamless data orchestration.
Analytical and Collaboration Skills
- Strong analytical and problem-solving skills with attention to detail.
- Effective communication and collaboration abilities to work with cross-functional teams, including product teams.
Nice-to-Have Qualifications
- Previous internship or project experience in data engineering.
Other Requirements
- Graduate in Dec 2026 or 2027 and pursuing a Bachelor's or Master's degree in Computer Science or a related field of study.
- Available for a 6 months full-time internship, starting on 18 May 2026.
Our Benefits:
Any general requests for consideration of your skills, please