Position Summary:
The Data Engineer is responsible for building and maintaining data pipelines ensuring the smooth operation of data systems and optimizing workflows to meet business requirements. This role will support data integration and processing for various applications.
Minimum Qualifications:
9 Years overall IT experience with minimum 4 years of work experience in below tech skills
Tech Skills:
Proficient in Python scripting and Spark for data processing tasks.
Strong SQL capabilities with handson experience managing big data using ETL tools like Informatica.
Experience with the AWS cloud platform and its data services including S3 Redshift Lambda EMR Airflow Postgres SNS and Event Bridge.
Skilled in BASH/Shell scripting.
Understanding of data Lakehouse architecture particularly with Iceberg format is a plus.
Preferred: Experience with Kafka and MuleSoft API.
Understanding of healthcare data systems is a plus.
Experience in Agile methodologies.
Strong analytical and problemsolving skills.
Effective communication and teamwork abilities.
Responsibilities:
Develop and maintain data pipelines and ETL processes to manage largescale datasets.
Collaborate to design & test data architectures to align with business needs.
Implement and optimize data models for efficient querying and reporting.
Assist in the development and maintenance of data quality checks and monitoring processes.
Support the creation of data solutions that enable analytical capabilities.
Contribute to aligning data architecture with overall organizational solutions.