Job Description
- Responsible for the maintenance of the company's big data business clusters/components, providing stable and efficient cluster services.
- Responsible for cluster capacity planning, expansion, and performance optimization.
- Design and implement operations/monitoring/management platforms for large-scale distributed clusters, improving work efficiency through automated operations tools.
- Participate in/lead infrastructure design, providing suggestions and solutions during the design phase.
- Stay updated with big data technology trends and research open-source technologies. Continuously optimize the cluster's foundational service architecture, apply new technologies to the big data platform/cluster, and drive team development.
Job Requirements
- Bachelor's degree or above in Computer Science, Computer Engineering, Information Systems, Business Analytics, Statistics, Mathematics or related fields.
- Proficient in at least one programming language in Python/Java/Scala, with solid programming fundamentals.
- Familiar with the Hadoop ecosystem, and well-versed in open-source projects such as Hadoop/HBase/Hive/Kafka/Zookeeper/Spark/Flink/ClickHouse.
- Knowledgeable about Kafka-based log streaming systems, and familiar with high-availability log systems like Filebeat and Logstash.
- Familiar with the configuration, management, and optimization of the Linux operating system, capable of independently troubleshooting and resolving OS-level issues. Strong learning ability with a keen interest in new technologies.
- Strong analytical and problem-solving skills, able to work under pressure in critical situations as well as day-to-day operations.
- Team player with positive and eager-to-learn attitude.
- Adaptable, open-minded, and comfortable to work with ambiguity and change in a fast-paced environment.
- Candidates must have at least 2 years of relevant work experience to be considered for the Engineer position, or 3 years for the Senior Engineer position.