Overview
The Data Engineer at TCS plays a crucial role in our data ecosystem. This position requires a deep understanding of data management technologies particularly in cloud environments like Azure and AWS. The Data Engineer works collaboratively with data architects data scientists and other stakeholders to design build and maintain scalable data pipelines and architectures that support analytics and business intelligence initiatives. Emphasizing automation and efficiency the Data Engineer is responsible for deploying data solutions that enhance data accessibility and usability across the organization. With a focus on Python PySpark and cloud platforms like Snowflake this role not only ensures high data quality but also supports our mission to leverage datadriven insights for strategic decisionmaking. The ideal candidate possesses technical acumen and a proactive mindset enabling them to adapt to evolving data technologies and improve data systems continuously.
Key Responsibilities
- Design and implement data models to support analytics and reporting efforts.
- Develop robust ETL processes for data ingestion from various sources.
- Utilize Python and PySpark for data processing and transformation.
- Create and manage cloudbased data warehouses on Azure and AWS.
- Optimize data architecture and storage solutions in Snowflake.
- Work closely with data scientists and business analysts to identify data requirements.
- Ensure data quality and integrity through rigorous validation processes.
- Implement data security measures to protect sensitive information.
- Automate data pipeline deployment and monitoring using best practices.
- Contribute to the development of data governance and data management strategies.
- Monitor and troubleshoot data pipeline issues to ensure reliability.
- Collaborate with crossfunctional teams to integrate data across platforms.
- Stay updated with the latest trends and technologies in data engineering.
- Document data processes and maintain technical specifications.
- Participate in code reviews and promote best coding practices.
Required Qualifications
- Bachelor s degree in Computer Science Engineering or related field.
- 3 years of experience as a Data Engineer or in a similar role.
- Proficient in Python and PySpark for data processing.
- Experience with Azure Data Factory Azure Synapse Analytics and related services.
- Handson experience with AWS Redshift and data lake solutions.
- Knowledge of Snowflake architecture and SQL querying.
- Strong understanding of data warehousing concepts and ETL frameworks.
- Familiarity with data modeling techniques and practices.
- Handson experience with Apache Spark and big data frameworks.
- Knowledge of data visualization tools is a plus.
- Experience with version control systems preferably Git.
- Ability to work collaboratively in a teamoriented environment.
- Exceptional problemsolving and analytical skills.
- Strong communication skills to convey technical concepts.
- Selfmotivated and eager to learn new technologies.
etl,aws,pyspark,data,sql,python,apache spark,data modeling,sql proficiency,aws redshift,azure,data warehousing,cloud computing,snowflake,data engineering