Bachelor’s degree or equivalent practical experience.
5 years of experience with software development in one or more programming languages, and with data structures/algorithms.
Experience in software development and engineering, incorporating design methodologies, leveraging open source technologies, and working with distributed computing systems, including Apache Spark, Apache Hadoop, and Apache Hive.
Experience in Open Source technologies, Big Data, Data Analytics, Artificial Intelligence, Machine Learning, and Database Internals.
Preferred qualifications:
Experience with database optimizations such as query and executor optimizations.
Experience with data lakes like Apache Iceberg, Apache Hudi, Delta Lake, etc.
Experience with Open Telemetry, JMX and other monitoring solutions.
Experience with OSS projects like Spark, Hive, Trino, Ray, Flink etc.
Experience working with data science tools such as Jupyter notebooks.