BS/MS in Computer Science or equivalent work experience
Spark / Spark Streaming hands-on, Scala & Java, Delta Lake (or Iceberg or Hudi): Basic understanding of Delta Lake concept
Working knowledge on Python, AWS (S3, EC2, EMR, IAM, Athena, Lambda), Kafka, Hive, Kubernetes (Spark on Kubernetes)
2-4years of overall experience developing systems/software for large business environments backend (Java/J2EE).
Strong experience leading design and implementation of robust and highly scalable web services
Skilled in software development lifecycle processes and experience with scrum, agile and iterative approaches
Expertise with unit testing & Test Driven Development (TDD)
Working knowledge of API development standards like REST/GraphQL protocol, pagination, throttling, caching and routing mechanisms experience is strongly preferred.
Strong understanding of Database, Data ingestion techniques
Experience building large scale data processing systems and solution that scales to PB volume.
Experience with Postgres, Data Lake, Kafka, & schema modelling.
Solid foundation in writing scalable web application code following best practices of accessibility, internationalization and instrumentation.
Good understanding of Software Design, Architecture and working with large scale systems.
An understanding of unit testing & Test-Driven Development (TDD)