Bachelor's degree in Computer Science, Mathematics, a related technical field, or equivalent practical experience.
3 years of experience in developing and troubleshooting data processing algorithms and software using Python, Java, Scala, Spark and hadoop frameworks.
Experience with distributed data processing frameworks and data stores.
Experience with SQL coding across databases (e.g., Teradata, MySQL).
Preferred qualifications:
Experience in Big Data, information retrieval, data mining, or Machine Learning with building applications with NoSQL, MongoDB, SparkML, and TensorFlow.
Experience architecting, developing software, internet scale production-grade Big Data solutions in environments.
Experience with data warehouses, including technical architectures, infrastructure components, ETL/ELT and reporting/analytic tools, environments, and data structures.
Experience with techniques like symmetric, asymmetric, Hardware Security Modules, envelopes, and implement secure key storage using Key Management System.
Experience with IaC and CI/CD tools like Terraform, Ansible, Jenkins etc.
Knowledge in data processing frameworks and data stores with ability to write SQLs.