Design, develop, and maintain cloud machine learning pipelines in production for real-time processing and batch processing.
Influence and craft the direction of a highly scalable machine learning pipeline.
Collaborate with data scientists to explore, design, and build pre/post-processing ETL pipelines.
Translate product requirements into a technical design and implement solutions using all resources available to you.
Build, monitor, and maintain CI/CD workflows for internal software/tools.
What we're looking for
Proficient and strong knowledge in Python, Terraform, cloud platforms (GCP/AWS), Docker, K8, Kafka, Github Actions, serverless, and SQL.
Familiar with Numpy, Tensorflow, Pytorch, vector DB, neural networks, computer vision, LLM, RAG and general ML concepts.
Rich experience in building large-scale streaming ETL pipelines, API endpoints, and serverless apps.
Proficiency in data manipulation and analysis using tools like SQL, BigQuery, Pandas, and/or Spark, along with experience in data visualization and dashboarding.
Strong interpersonal and communication skills, with successful experience collaborating with distributed teams to deliver complex solutions.
You might also have
Experience with distributed computing libraries such as Ray or Dask.
Experience in A/B testing and statistical analysis.
Experience with ML monitoring solutions (Prometheus, Grafana, etc) in a public cloud environment (GCP, AWS, Azure).
Additional information
Relocation support is not available for this position
International relocation support is not available for this position