Hadoop Bigdata Manager
Full Time Phnom Penh Posted 3 weeks ago
Technology
JOB RESPONSIBILITIES
- Provide technical support & analysis
- Architect, deploy, and manage HADOOP clusters using Cloudera Data Platform or Hortonworks Data Platform.
- Optimize HDFS, YARN, Hive, HBase, Spark, and Kafka for high-performance data processing
- Managing Cloudera Manager or Ambari for cluster monitoring, configuration, and troubleshooting.
- Implement data replication and streaming solutions using Oracle Goldengate, Qlik Replicate, or Kafka to ensure high availability.
- Design and integrate Lakehouse architecture for scalable, high performance analytic.
- Perform capacity planning, cluster scaling, and performance tuning to optimize resource utilization
- Automate administrative tasks using Python, Shell script, or terraform for infrastructure management
- Continuously assess and refine data workflows to improve efficiency and reduce time-to-insight
- Establish and enforce data governance policies to ensure compliance with industry standard.
- Oversea the implementation of security protocols (e.g., Kerberos, Apache Ranger, and Sentry) to protect sensitive enterprise data.
- Stay ahead of emerging big data technologies and drive innovation within the organization.
- Collaborate with IT, business stakeholders, and data scientist to prioritize projects and drive data-driven decision-making
- Set a clear performance metric, provide feedback, and foster a culture of innovation within the team.
JOB REQUIREMENTS
- Graduated bachelor degree of Information Technology, preferably in the field of Computer Science.
- Bachelor’s or Master’s degree in computer science, Data Engineering, or related field.
- 4+ year of experience in Hadoop administration, big data management, or related fields.
- Strong expertise in Hadoop ecosystem (HDFS, YARN, Hive, HBase, Spark, Nifi, Kafka)
- Hands-on experience with Cloudera Data Platform (CDP) or Hortonworks Data Platform (HDP).
- Proficiency in Cloudera Manager, Ambari, Ranger and Sentry for cluster management and security.
- Experience with Lakehouse technologies Apache iceberg, Hudi, particularly Dremio for high-performance analytic.
- Strong understanding of data governance, compliance, and security best practices.
- Familiarity with monitoring tools like Grafana, Prometheus, Cloudera Manager, or Ambari.
- Demonstrate ability to lead diverse technical teams and manage complex projects.
- Strong interpersonal and communication skills to effectively liaise with technical team and business stakeholders.