Essential Responsibilities:
- Develop and optimize machine learning models for various applications.
- Preprocess and analyze large datasets to extract meaningful insights.
- Deploy ML solutions into production environments using appropriate tools and frameworks.
- Collaborate with cross-functional teams to integrate ML models into products and services.
- Monitor and evaluate the performance of deployed models.
Minimum Qualifications:
- Minimum of 5 years of relevant work experience and a Bachelor's degree or equivalent experience.
- Experience with ML frameworks like TensorFlow, PyTorch, or scikit-learn.
- Familiarity with cloud platforms (AWS, Azure, GCP) and tools for data processing and model deployment.
- Several years of experience in designing, implementing, and deploying machine learning models.
Preferred Qualification:
- Strong programming skills in Big data processing(Pig/Scala+Java/Python)and SQL.
- Strong in data issue investigation and problem solving. Ability to synthesize information and generalize the pattern.
- Expertise on big data platform and infrastructure
- Develop, optimize and maintain ETL pipelines to handle large volumes of data from multiple sources for advanced machine learning models
- Build and optimize distributed data processing systems using big data frameworks and technologies
- Maintain and improve existing data infrastructure, ensuring high availability and fault tolerance
- Collaborate with engineers from other sites, data scientists and business stakeholders to understand data requirements and deliver appropriate solutions
- Strong proficiency in Python, Java or Scala
- Extensive experience with Apache Spark (Spark SQL, Spark Streaming, PySpark)
- Hands-on experience with Hadoop ecosystem (HDFS, YARN, Hive, HBase)
- Experience with cloud-based data platforms (Google BigQuery)
- Experience with relational databases (e.g., PostgreSQL, MySQL) and/or NoSQL databases (e.g., MongoDB)
- Experience with version control systems (Git) and CI/CD practices
- Familiar with Linux environments; able to perform troubleshooting and write automation scripts (Shell/Python)
- Good documentation habit and can sync up with teams across different locations remotely
- Good understanding of security principles and data protection
- Experience with time-series databases (InfluxDB)
- Knowledge of RESTful API development and HTTP client libraries
- Experience in building GenAI based solutions
- Strong problem-solving skills and attention to detail
- Experience working in agile development environments
- Excellent communication and collaboration skills
Our Benefits:
Any general requests for consideration of your skills, please