About Client:
Client is a global leader in professional services and solutions at the forefront of AI data engineering and cloud technologies. With their distributed team of over 500 analytics professionals across North America Europe and Asia they are dedicated to enabling global enterprises to transform and innovate through the power of AI (Artificial Intelligence) data engineering and cloud technologies.
They offer exciting career opportunities for professionals seeking to make a significant impact. Their diverse team represents a rich tapestry of talent expertise and cultural backgrounds.
Job Description
Notice Period: : Upto 30 Days
Experience: 815 Years
Roles and Responsibilities:
Seeking an experienced Data Engineer. The ideal candidate will have 8 to 12 years of experience in data engineering and a strong background in building data pipelines data warehousing and data management solutions. In this role you will work closely with data scientists analysts and other stakeholders to design build and maintain scalable and robust data infrastructure.
- Design build and operationalize large scale enterprise data solutions and applications using one or more of AWS data and analytics services Glue RedShift Lambda S3.
- Ability to create pyspark/python code from scratch according to the business requirement while implementing the industry standard best practices.
- Hands on experience utilizing AWS Management Tools (CloudWatch CloudTrail) to proactively monitor large and complex deployments.
- Experience in performing code version management through CodeCommit.
- Experience in analyzing rearchitecting and replatforming onpremise data warehouses to data platforms on AWS.
- Excellent in Advanced SQL end to end development and Optimization of existing processes
- Table Partition mechanism Types of Partitions column store index
- Maintenance and optimization of existing processes
- Requirement gathering and feasibility analysis design develop from scratch and deliver
- Experience in designing data warehouse/data mart.
- Leading the client calls to flag off any delays blockers escalations and collate all the requirements
- Experience in deploying data pipelines using automated CI/CD approach.
- Experience in writing production ready code in Python and test participate in code reviews to maintain and improve code quality stability and supportability.
- Ability to write modularized reusable code components.
- Proficient in identifying data issues and anomalies during analysis.
- Leading the client calls to flag off any delays blockers escalations and collate all the requirements
Must Have
Experience in Data Engineering (Min experience of 8 Years)
Experience in Python OR Pyspark (Either of them is Mandate) Min. 67 Years experience
Exposure in Data Warehousing and Data Migration
Experience with AWS services (67 years) (DynamoDB RedShift Lambda and Glue) out of these services candidate should have experience with at least 2 of these service.
SQL experience
aws,python,sql,data migration,cloud,data warehousing