Job Responsibilities:
- Develop and maintain robust ETL processes for data integration.
- Utilize big data technologies like Apache Spark (PySpark) for data processing.
- Write secure and high-quality code in Python and SQL.
- Build cloud-native applications using platforms such as AWS, Azure, or GCP (AWS).
- Leverage cloud services for data storage, processing, and analytics.
- Implement containerization technologies like Docker and Kubernetes (EKS).
- Apply object-oriented design and data structure fundamentals.
- Collaborate effectively with cross-functional teams.
Required Qualifications, Capabilities, and Skills:
- Formal training or certification on software engineering concepts and 2+ years applied experience
- Strong hands-on experience in developing and maintaining ETL processes.
- Knowledge of big data technologies such as Apache Spark (PySpark).
- Proficiency in Python and SQL.
- Experience with cloud platforms like AWS, Azure, or GCP (AWS).
- Hands-on experience with Docker and Kubernetes (EKS).
- Solid understanding of object-oriented design and data structures.
- Strong collaboration skills.
Preferred Qualifications, Capabilities, and Skills:
- Experience in using Databricks for big data analytics.
- Experience with data orchestrator tools like Airflow or Prefect.