This is a remote position.
We are seeking a skilled Data Engineer with a strong background in Google Cloud Platform (GCP) services and experience in creating and managing data pipelines. The ideal candidate will have GCP Data Engineer certification and be proficient in various GCP tools and services for data orchestration transformation and reporting.
Key Responsibilities:
- Utilize Cloud Scheduler for scheduling queries Cloud Run Cloud Functions BigQuery and other GCP services to manage data workflows and processes.
- Implement data pipelines using Dataproc with PySpark for efficient data processing.
- Use Cloud Composer to orchestrate and manage data pipelines across the platform.
- Perform data modeling and transformation using DBT to structure and prepare data for analysis.
- Deploy infrastructure as code with Terraform for a reliable and reproducible environment setup.
- Leverage Power BI for reporting and visualization to meet business intelligence needs.
Requirements
- GCP Data Engineer certification
- Proficiency with core GCP services including BigQuery Cloud Scheduler Cloud Run Cloud Functions Dataproc (PySpark) and Cloud Composer.
- Experience with DBT for data transformation and Terraform for IaC deployments.
- Knowledge of Power BI for reporting.
Preferred:
- Familiarity with Azure and related services (not mandatory).