Roles & Responsilbities
- Strong ability to build and maintain reliable and scalable data pipeline using Python or any other ETL tool (Pentaho Data Integration preferred)
- Identify, design, and implement internal process improvements: automating manual processes optimizing data delivery, and re-designing infrastructure for greater scalability.
- Strong SQL Querying skills (MySQL, SQL Server, and MongoDB)
- Strong understanding of Data Warehousing concepts
- Skilled in Data modeling and Dimensional modeling techniques
- Experience with AWS technologies (EC2, RDS, S3, Lambda, Elasticsearch, Athena, etc.)
- Recommend different ways to constantly improve data reliability and quality
- Possess an ability to perform complex data analyses with large data volumes
- Strong understanding & usage of algorithms and data structures.
Nice to Have:
- AWS Redshift knowledge and operational experience
- Hands on experience in real-time Big Data projects using Distributed data pipelines (EMR, Hadoop, Spark)and/or streaming pipeline frameworks (Kinesis, Kafka, Spark)
- Machine Learning expertise
- Experience with analytical or data visualization tools like Tableau
Skills : ETL Tool,Python,SQL,Data Warehousing