Responsible for the maintenance of the company’s big data business clusters/components, providing stable and efficient cluster services.
Responsible for cluster capacity planning, expansion, and performance optimization.
Design and implement operations/monitoring/management platforms for large-scale distributed clusters, improving work efficiency through automated operations tools.
Participate in/lead infrastructure design, providing suggestions/solutions during the design phase.
Stay updated with big data technology trends, research open-source technologies, continuously optimize the cluster’s foundational service architecture, apply new technologies to the big data platform/cluster, and drive team development.
Requirements & Skills:
Bachelor’s degree or above in Computer Science, Computer Engineering, Information Systems, Business Analytics, Statistics, Mathematics or related field
Proficient in at least one programming language in Python/Java/Scala, with solid programming fundamentals.
Familiar with the Hadoop ecosystem, and well-versed in open-source projects such as Hadoop/HBase/Hive/Kafka/Zookeeper/Spark/Flink/ClickHouse.
Knowledgeable about Kafka-based log streaming systems, and familiar with high-availability log systems like Filebeat and Logstash.
Familiar with the configuration, management, and optimization of the Linux operating system, capable of independently troubleshooting and resolving OS-level issues. Strong learning ability with a keen interest in new technologies.
Strong analytical and problem-solving skills, in critical situations under pressure as well as day-to-day operations.
Team players with positive and eager-to-learn attitude
Adaptable, open-minded, and comfortable to work with ambiguity and change in a fast-paced environment
Candidates must have at least 2 years of relevant work experience for the Engineer position or 3 years for the Senior Engineer position.