Experience: 3.5 12 Years
Location: Hyderabad
Mandatory Skills: Python/ PySpark / Spark with Azure/ Databricks / Azure Data Factory
Primary Roles and Responsibilities:
Developing Modern Data Warehouse solutions using Databricks and AWS/ Azure Stack
Ability to provide solutions that are forwardthinking in data engineering and analytics space
Collaborate with DW/BI leads to understand new ETL pipeline development requirements.
Triage issues to find gaps in existing pipelines and fix the issues
Work with business to understand the need in reporting layer and develop data model to fulfill reporting needs
Help joiner team members to resolve issues and technical challenges.
Drive technical discussion with client architect and team members
Orchestrate the data pipelines in scheduler via Airflow
Skills and Qualifications:
Bachelors and/or master s degree in computer science or equivalent experience.
Must have 5 years experience in Data warehouse/ETL projects.
Deep understanding of Star and Snowflake dimensional modelling.
Strong knowledge of Data Management principles
Good understanding of Databricks Data & AI platform and Databricks Delta Lake Architecture
Should have handson experience in SQL Python and Spark (PySpark)
Candidate must have experience in AWS/ Azure stack
Desirable to have ETL with batch and streaming (Kinesis).
Experience in building ETL / data warehouse transformation processes
Experience with Apache Kafka for use with streaming data / eventbased data
Experience with other OpenSource big data products Hadoop (incl. Hive Pig Impala)
Experience with Open Source nonrelational / NoSQL data repositories (incl. MongoDB Cassandra Neo4J)
Experience working with structured and unstructured data including imaging & geospatial data.
Experience working in a Dev/Ops environment with tools such as Terraform CircleCI GIT.
Proficiency in RDBMS complex SQL PL/SQL Unix Shell Scripting performance tuning and troubleshoot
Databricks Certified Data Engineer Associate/Professional Certification (Desirable).
Comfortable working in a dynamic fastpaced innovative environment with several ongoing concurrent projects
Should have experience working in Agile methodology
Strong verbal and written communication skills.
Strong analytical and problemsolving skills with a high attention to detail.Tips: Provide a summary of the role Data bricks what success in the position looks like and how this role fits into the organization overall.
etl,pipeline,aws,pl/sql,azure,data,design,nosql,unix shell scripting,sql,microsoft azure,skills,data warehouse,azure data factory,dbx,circleci,apache kafka,dimensional modelling,data management,spark,terraform,adfs,pyspark,datasynapse gridserver,databricks,python,azure functions,hadoop,rdbms,git