About Company:
We are hiring for a client that thrive on the dynamic energy of a startup environment while harnessing the power of cuttingedge technologies to drive innovation in the digital landscape. As a forwardthinking organization we are dedicated to pushing the boundaries of possibility delivering impactful solutions that transform industries and empower businesses to achieve their full potential. Join our team of passionate individuals who are committed to shaping the future of datadriven decisionmaking.
Responsibilities:
- Manage all aspects of data extraction transfer and loading activities in a fastpaced startup setting.
- Develop robust data pipelines to ensure seamless data access across platforms.
- Execute ETL processes including data ingestion cleaning and curation into our data warehouse or platform.
- Contribute to data modeling and ML pipelines within the AI/ML ecosystem.
- Collaborate with DevOps and senior Architect teams to design scalable system and model architectures for realtime and batch services.
Requirements:
- Minimum 5 years of experience in data engineering or data science with a focus on data engineering and ETL jobs.
- Deep understanding of data warehousing data modeling and data analysis concepts.
- Proficiency in building and managing pipelines and performing ETL using Redshift (2 years of experience).
- Ability to troubleshoot and resolve performance issues on Redshift.
- Familiarity with orchestration tools like Airflow along with strong Python and SQL coding skills.
- Experience with distributed systems like Spark and AWS Data and ML Technologies.
- Handson experience with data extraction techniques and related tools for near realtime and batch data extraction.
Note:
Candidates with experience in productbased companies are preferred.
data engineering,data analytics,etl,spark,aws,data extraction,data loading,data cleaning,data science,python,sql,distributed systems,cdc,kafka,glue,athena,aws lambda,airflow,amazon redshift,orchestration,ml,pipelines