PLEASE NOTE THIS IS AN EQUITYONLY ROLE AND THE INTERVIEWS WILL COMMENCE IN FEBRUARY 2025.
StealthMode StartUp Client is seeking skilled Data Engineers to design build and maintain scalable data pipelines and infrastructure that power advanced analytics AI and machine learning initiatives across the platform. This role will focus on data ingestion storage processing and integration ensuring that data flows efficiently across systems and remains secure reliable and readily available for analysis.
The ideal candidate will have expertise in big data technologies cloud infrastructure and data architecture design with a passion for enabling datadriven decisions at scale.
To apply please provide a CV your compensation requirements (including salary expectations for when funding is secured) and a cover letter/note that explains why you are interested and how you meet the requirements. Please note that submissions received without all the requested information will be automatically disqualified and rejected.
Key Responsibilities:
- Design and implement scalable ETL/ELT data pipelines to collect process and store large datasets from diverse sources.
- Develop and maintain a robust data architecture to support realtime and batch data processing workflows.
- Integrate data from multiple platforms APIs and thirdparty services into centralized data lakes and warehouses.
- Optimize and manage cloudbased data infrastructure (e.g. AWS Redshift GCP BigQuery Azure Synapse).
- Implement data validation and integrity checks to ensure data accuracy completeness and consistency across systems.
- Build and manage realtime data streaming pipelines using technologies like Apache Kafka Spark Streaming or similar.
- Work closely with Data Scientists Analysts and Product Managers to understand data requirements and ensure data availability for analytical models.
- Continuously optimize data storage and retrieval processes for performance and cost efficiency.
- Implement data encryption access control and compliance with regulatory standards (e.g. GDPR CCPA).
- Set up monitoring systems and logging mechanisms to track data flow detect anomalies and resolve pipeline failures.
- Maintain detailed technical documentation for all data pipelines workflows and infrastructure components.
Requirements:
- Minimum 3 years of experience as a Data Engineer or in a similar role.
- Excellent command of the English Language in all forms.
- Previous startup experience would be an advantage.
- Proficiency with ETL/ELT tools (e.g. Apache Airflow Talend dbt).
- Strong experience in languages like Python Java or Scala for data processing and automation.
- Experience with Hadoop Spark Kafka or similar big data technologies.
- Proficiency with SQL and NoSQL databases (e.g. PostgreSQL MongoDB Cassandra).
- Handson experience with cloud data platforms such as AWS Redshift GCP BigQuery or Azure Data Lake.
- Familiarity with data security best practices and compliance with global privacy regulations (e.g. GDPR).
- Experience with monitoring and logging platforms (e.g. Prometheus Datadog Splunk).
- Proven ability to design systems that handle largescale data volumes efficiently.
- Strong analytical and problemsolving skills to troubleshoot complex data engineering challenges.
- Ability to work crossfunctionally with Data Scientists Analysts Product and Engineering Teams.
Ideal Candidate Profile:
- A detailoriented engineer passionate about building scalable data systems from scratch.
- Thrives on solving complex data integration challenges across diverse datasets and platforms.
- Proactive in identifying bottlenecks and optimizing workflows for better efficiency.
- Collaborative with excellent communication skills for sharing technical details with nontechnical stakeholders.
- Continuously curious about emerging data technologies and tools to improve data processing capabilities.
- Committed to maintaining data integrity security and compliance across all systems.
Compensation & Benefits
Equityonly at present to transition to a salaried fulltime permanent position when funding is secured.
Remote and flexible working arrangements the opportunity to be part of something potentially epic with potential opportunities for global travel and access to industry conferences and workshops in due course.