Overview
The Data Warehouse Engineer plays a critical role in managing and implementing data solutions and infrastructure for our organization. This position focuses on the design development and maintenance of data warehouse solutions that enable the company to turn raw data into actionable insights. By ensuring that the data is accessible reliable and optimized for reporting and analytics the Data Warehouse Engineer supports various departments in datadriven decisionmaking. In a fastpaced environment where data is continually generated this position is essential for creating robust frameworks and refining existing processes for data handling. The Data Warehouse Engineer will collaborate with data analysts business intelligence teams and IT professionals to ensure that the organization has a strong data foundation enabling effective reporting trend analysis and strategic planning.
Key Responsibilities
- According to the companys data warehouse specifications and business understanding build a universal and flexible data warehouse system that can quickly support the needs and reduce repetitive development work efforts.
- Data model design development testing deployment online data job monitoring and the ability to quickly solve complex problems especially the optimization of complex calculation logic and performance tuning etc.
- Participate in Data governance including the construction of the company s metadata management system and data quality monitoring system.
- Participate in technical team building and learning growth and contribute to the team s overall knowledge accumulation and skill improvement.
Required Qualifications
- 6 years experiences of data lake and data warehouse design and development experience.
- Deeply understanding of data warehouse modeling and data governance.
- Solid knowledge of data warehouse development methodology including dimensional modeling information factory and one data etc.
- Proficient in Java / Scala / Python (at least one language) and Hive & Spark SQL programming languages.
- Familiar with OLAP technology (such as: kylin impala presto druid etc.).
- Proficient in Big Data batch pipeline development.
- Familiar with Big Data components including but not limited to Hadoop Hive Spark Delta lake Hudi Presto Hbase Kafka Zookeeper Airflow Elasticsearch Redis etc.
- Experiences with AWS Big Data services are a plus.
- Clear mind with good business requirement understanding analysis abstraction and system design capabilities.
- Those who have experienced data volume above PB level in Internet companies and solve difficult production issues are preferred.
hive,delta lake,dimensional modeling,scala,data,data lake,data modeling,presto,zookeeper,hadoop,aws big data services,data warehousing,hudi,design,data governance,java,airflow,spark sql,python,kafka,data warehouse design,olap technology,olap,big data,elasticsearch,redis,hbase