We’re working on a large scale of Data, creating and maintaining Data-Lake infrastructure:
* Based on AWS services.
* Orchestrating with Airflow
* Developing automatic ETL tools.
* Processing Presto/Trino over EMR.
Incharge of ~1.5B events per day, including log-servers written in Go-lang.
Working with realtime and batch data.
Using Terraform/Docker as an infrastructure
Maintaining BigData databases like druid, Singlestore and others.
Designing and implementing API on top of our data using GraphQL, written in Go-lang.
What you'll be doing
- Write Airflow Dags using Python to pull data from different API's (Facebook, google, tiktok etc..)
- Develop automatic ETL tools
- Implementing API on top of the data using GraphQL, written in Go.
- Create and maintains the infrastructure for distributed data processing (EMR clusters)
What we're looking for
- Proficient in programming languages such as Python, Go.
- Strong knowledge of SQL.
- Hive, Presto/Spark skills are highly important.
- Good knowledge of Unix/Linux environment & Docker / Terraform
- B.A/B.Sc degree in Computer Science /industrial engineering /military service in a relevant technological unit
You might also have
- NoSQL (Redis, DynamoDB, MongoDB) DB
- Experience with API GraphQL
- Knowledge with Snowflake
Additional information
- Relocation support is not available for this position.
- International relocation support is not available for this position.
- Work visa/immigration sponsorship is not available for this position.