Overview:
The Data Engineer plays a crucial role in the organizations data infrastructure and is responsible for developing constructing testing and maintaining architectures such as databases and largescale processing systems. This role requires a keen understanding of data systems strong analytical and problemsolving skills and the ability to work closely with crossfunctional teams to ensure seamless data integration and optimal data delivery for diverse projects.
Key Responsibilities:
- Design build and maintain scalable data pipelines and ETL processes for diverse data sets.
- Collaborate with data scientists and analysts to understand their data processing needs and develop appropriate solutions.
- Implement best practices for data security and privacy throughout the data infrastructure.
- Optimize and tune existing data processes to improve performance and reliability.
- Work with stakeholders to understand data requirements and translate them into technical specifications.
- Develop and maintain documentation for data infrastructure and processes.
- Participate in the evaluation and implementation of new tools and technologies for data management and analytics.
- Collaborate with crossfunctional teams to ensure data compliance consistency and quality.
- Contribute to the development of data governance and quality standards.
- Perform data modeling schema design and implementation.
- Monitor data infrastructure to ensure efficient and reliable operation.
- Troubleshoot and resolve data infrastructure and pipeline issues.
- Conduct performance tuning and optimization of data systems.
- Implement data solutions for realtime and batch processing.
Required Qualifications:
- Bachelors degree in Computer Science Information Technology or related field.
- Proven experience in data engineering or related roles.
- Proficiency in programming languages such as Python Java or Scala.
- Expertise in SQL and database technologies (e.g. PostgreSQL MySQL or similar).
- Strong understanding of data modeling and architectural patterns.
- Experience with ETL processes and tools.
- Knowledge of big data technologies and frameworks such as Hadoop Spark or similar.
- Familiarity with cloud platforms and services (e.g. AWS Azure or Google Cloud Platform).
- Ability to collaborate effectively with crossfunctional teams.
- Excellent problemsolving and analytical abilities.
- Strong communication and documentation skills.
- Experience with data governance and privacy regulations is a plus.
- Ability to work in a fastpaced and dynamic environment.
- Understanding of machine learning and data analytics concepts is desirable.
- Continuous learning and staying updated on industry trends and developments.
etl processes,data governance,python,data analytics,machine learning,big data technologies,java,data modeling,spark,aws,hadoop,cloud platforms,data engineering,sql,scala,google cloud platform,big data,etl,azure