Key Responsibilities:
1. Install configure and manage Hadoop clusters (HDFS MapReduce YARN etc.)
2. Monitor cluster performance identify bottlenecks and optimize resource utilization.
3. Implement and manage Hadoop security (Kerberos ACLs encryption).
4. Develop and maintain scripts for automation (Python Shell).
5. Collaborate with developers to optimize data processing workflows.
6. Troubleshoot Hadooprelated issues and provide Level 3 support.
7. Plan and execute cluster upgrades patches and backups.
8. Ensure data quality integrity and compliance with organizational policies.
9. Document cluster architecture configuration and procedures.
10. Work closely with data engineers analysts and scientists to meet business requirements.
Requirements:
Technical Skills:
1. 3 years of experience in Hadoop administration.
2. Strong understanding of Hadoop ecosystem (HDFS MapReduce YARN Hive Pig Sqoop Flume).
3. Experience with cluster management tools (Ambari Cloudera Manager).
4. Proficient in Linux/Unix and scripting languages (Python Shell).
5. Knowledge of Hadoop security and authentication protocols.
6. Familiarity with data processing frameworks (Spark Flink).
7. Experience with cloudbased Hadoop deployments (AWS Azure GCP).