Design new processes and builds large complex data flows can possess excellent analytical skills and troubleshooting ideas aware for Agile Mode of operationsopen to work in Devops model.
Responsibilities:
- Design new processes and builds large complex data flows
- Should possess excellent analytical skills and troubleshooting ideas
- Should be aware for Agile Mode of operations and should have been part of scrum teams.
- Should be open to work in Devops model with responsibilities of Dev and Support both as application goes live
- Should be able to work in shifts (if required)
- Should be open to work in fast paced project with multiple stakeholders.
Key Skills:
- 3 years of experience
- Bachelors/Masters Degree in Computer Science IT or similar
- Databricks Delta lake experience
- Experience building data pipelines using ADF
- Strong knowledge in R/Python/Rubys
- Advanced proficiency in Java/Scala SQL NoSQL
- strong in Database and Data Warehousing concepts
- Expertise in SQL tuning schema design Python and ETL processes
- Highly Motivated Selfstarter and quick learner
- Proficiency in Statistical procedures Experiments and Machine Learning techniques
- Must have knowledge on basics of data analytics and data modelling.