Roles and responsibilities
- Liaising with coworkers and clients to elucidate the requirements for each task.
- Conceptualizing and generating infrastructure that allows big data to be accessed and analyzed.
- Reformulating existing frameworks to optimize their functioning.
- Testing such structures to ensure that they are fit for use.
- Preparing raw data for manipulation by data scientists.
- Detecting and correcting errors in your work.
- Ensuring that your work remains backed up and readily accessible to relevant coworkers.
- Remaining up-to-date with industry standards and technological advancements that will improve the quality of your outputs.
-
• Gathering the ETL/ELT requirements.
• Design, Develop, Test ETL/ELT different scenarios and support the ETL/ELT processes to Extract, Transforming and Loading data from different sources to the data warehouse.
• Create and maintain optimal integration solutions including data pipeline architecture.
• Analyze multiple sources of structured and unstructured data to propose and design data architecture solutions for scalability, high availability, fault tolerance, and elasticity.
• Assist architects in the design and implementation of high-performance systems with large volume data integration processes, database, storage and provisioning.
- Meeting with managers to determine the company’s Big Data needs.
- Developing Hadoop systems.
- Loading disparate data sets and conducting pre-processing services using Hive or Pig.
- Finalizing the scope of the system and delivering Big Data solutions.
- Managing the communications between the internal system and the survey vendor.
- Collaborating with the software research and development teams.
- Building cloud platforms for the development of company applications.
- Maintaining production systems.
- Training staff on data resource management.
Desired candidate profile
- Strong communication, interpersonal, and presentation skills.
- Ability to multi-task and work under pressure.
- Excellent command of English.
- Excellent verbal and written communication skills; influencing skills and ability to work effectively in a geographically dispersed team.
- Bachelor's degree in data engineering, big data analytics, computer engineering, or related field.
- Master's degree in a relevant field is advantageous.
- Proven experience as a data engineer, software developer, or similar.
- Expert proficiency in Python, C++, Java, R, and SQL.
- Familiarity with Hadoop or suitable equivalent.
- Excellent analytical and problem-solving skills.
- A knack for independence and group work.
- Scrupulous approach to duties.
- Capacity to successfully manage a pipeline of duties with minimal supervision.