Title: Data Engineer
Location: Bangalore Karnataka
Timings: Full Time (As per company timings)
Notice Period: within 15 days or immediate joiner
Experience: 4 Years
Interview round 1st round Assignment round Final F2F round.
Basic Qualifications
- Bachelors degree in Computer Science Information Technology or a related field.
- Prior professional noninternship experience in software development specifically within product startup environments.
- Proven experience as a Data Engineer with expertise in ETL techniques is a MUST
- Strong programming skills in languages such as Python Java or Scala.
- Skillset to se and transform data off of the publicly available web sources
- Experience with cloudbased data platforms (e.g. AWS Azure GCP).
- Proficiency in SQL and experience working with relational and nonrelational databases.
- Knowledge of data warehousing concepts and architectures.
- Familiarity with big data technologies such as Hadoop Spark and Kafka.
- Experience with data modeling tools and techniques.
- Excellent problemsolving and analytical skills.
- Strong communication and collaboration skills.
Preferred Qualifications
- Masters degree or equivalent in Computer Science/ Data Science
- Knowledge of data streaming and realtime processing.
- Familiarity with data governance and security best practices.
Responsibilities:
Outline the primary responsibilities and tasks the candidate is expected to perform.
About the role
As a Data Engineer for our Data Science team you will play a critical role in helping the team to enrich and maintain the central repository of datasets which is leveraged for carrying out advanced data analytics and machine learning techniques to extract actionable insights from financial and market data. You will work closely with crossfunctional teams to develop and implement robust data pipelines that will automate the updation of data in our cloudbased repository in a ready to use form thereby increasing data accessibility for the entire organization.
Key Responsibilities:
ETL Development:
- Design develop and maintain efficient ETL processes for multiscale data sets.
- Implement and optimize data transformation and validation processes to ensure data accuracy and consistency.
- Collaborate with crossfunctional teams to understand data requirements and business logic.
Data Pipeline Architecture:
- Architect build and maintain scalable and highperformance data pipelines
- Evaluate and implement new technologies to enhance data pipeline efficiency and reliability
- Pipelines for extracting data through sing for adhoc sector specific datasets
Data Modelling:
- Develop and implement data models to support analytics and reporting needs.
- Optimize database structures for performance and scalability.
Data Quality and Governance:
- Implement data quality checks and governance processes to ensure data integrity.
- Collaborate with stakeholders to define and enforce data quality standards.
Documentation and Communication:
- Document ETL processes data models and other relevant information.
- Communicate complex technical concepts to nontechnical stakeholders effectively.
Crossfunctional collaboration:
- Collaborate internally with the Quant team and developers to lay and optimize the data pipelines and externally with the stakeholders to understand the business requirements for the enrichment of the cloud database