Hadoop Bigdata Manager

Full Time Phnom Penh Posted 3 weeks ago

Technology

JOB RESPONSIBILITIES

  • Provide technical support & analysis
  • Architect, deploy, and manage HADOOP clusters using Cloudera Data Platform or Hortonworks Data Platform.
  • Optimize HDFS, YARN, Hive, HBase, Spark, and Kafka for high-performance data processing
  • Managing Cloudera Manager or Ambari for cluster monitoring, configuration, and troubleshooting.
  • Implement data replication and streaming solutions using Oracle Goldengate, Qlik Replicate, or Kafka to ensure high availability.
  • Design and integrate Lakehouse architecture for scalable, high performance analytic.
  • Perform capacity planning, cluster scaling, and performance tuning to optimize resource utilization
  • Automate administrative tasks using Python, Shell script, or terraform for infrastructure management
  • Continuously assess and refine data workflows to improve efficiency and reduce time-to-insight
  • Establish and enforce data governance policies to ensure compliance with industry standard.
  • Oversea the implementation of security protocols (e.g., Kerberos, Apache Ranger, and Sentry) to protect sensitive enterprise data.
  • Stay ahead of emerging big data technologies and drive innovation within the organization.
  • Collaborate with IT, business stakeholders, and data scientist to prioritize projects and drive data-driven decision-making
  • Set a clear performance metric, provide feedback, and foster a culture of innovation within the team.

JOB REQUIREMENTS

  • Graduated bachelor degree of Information Technology, preferably in the field of Computer Science.
  • Bachelor’s or Master’s degree in computer science, Data Engineering, or related field.
  • 4+ year of experience in Hadoop administration, big data management, or related fields.
  • Strong expertise in Hadoop ecosystem (HDFS, YARN, Hive, HBase, Spark, Nifi, Kafka)
  • Hands-on experience with Cloudera Data Platform (CDP) or Hortonworks Data Platform (HDP).
  • Proficiency in Cloudera Manager, Ambari, Ranger and Sentry for cluster management and security.
  • Experience with Lakehouse technologies Apache iceberg, Hudi, particularly Dremio for high-performance analytic.
  • Strong understanding of data governance, compliance, and security best practices.
  • Familiarity with monitoring tools like Grafana, Prometheus, Cloudera Manager, or Ambari.
  • Demonstrate ability to lead diverse technical teams and manage complex projects.
  • Strong interpersonal and communication skills to effectively liaise with technical team and business stakeholders.
Apply