Job responsibilities:
- Implementing Data Ingestion pipelines for Batch and streaming data at petabyte scale
- Experience in building and deploying machine learning / AI models in the cloud
- Run machine learning tests / experiments and perform statistical analysis and finetuning using test results
- To research experiment with and implement suitable ML algorithms and tools.
- Work with internal and external stakeholders to assist with datarelated technical issues and support data infrastructure needs
- Optimize data delivery and redesign infrastructure for greater scalability
- Identify design and implement internal process improvements
Preferred Skills & Qualifications:
- Computer Science / Computer Engineering Degree with Minimum of 8 years data Engineering experience including 3 years of MLOps experience.
- Experience in requirement gathering analysis and solution design.
- Proficient in Python and strong in SQL.
- Experience in data and ML pipeline implementation using Python/R/SQL/Docker/Kubernetes/Domino Data Lab .
- Experience in Snowflake AWS and Kafka.
- Experience in Scheduling tools like ControlM and Airflow.
- Experience in Agile development methodologies.
- Deep knowledge of math probability statistics and algorithms
- Excellent algorithm and data structure skills
- Strong problem solving and troubleshooting skills with the ability to exercise mature judgment.
- Excellent interpersonal and collaboration skills
Additional skills :
- Experience in GenAI and AWS sagemaker is a plus.
- Certified in Snowflake and AWS is a plus.
- Experience in Financial Domain is a plus.