Location:
CN-Shenzhen-HyQ
Shift:
Standard - 40 Hours (China)
Scheduled Weekly Hours:
40
Worker Type:
Permanent
Job Summary:
The AVP under Site Reliability Engineering (SRE) in the Chief Data Office plays a critical role in maintaining and enhancing the reliability performance and scalability of our enterprise data platform. This position ensures the stability of production testing and development environments that support Big Data and Analytics initiatives. You will leverage expertise in DevOps practices Infrastructure as Code and Kubernetes Architecture to drive automation streamline operations and proactively resolve system issues. A strong background in managing cloud-based large-scale data infrastructure is essential along with advanced Python scripting skills for automation and infrastructure management.
Job Duties:
Job Duties
- Build and maintain cloud-based Big Data and Analytics Platforms including:
- Enterprise Data Lake
- Data Governance and Management Platforms
- Self-service BI and Augmented Analytics Platforms
- Proactively managing production services and data pipelines to ensure availability and system health
- Improve system reliability through engineering solutions and advanced tooling for monitoring automation and fault tolerance
- Develop and maintain CI/CD tools microservices and solutions to meet customer service standards and minimize errors
- Manage release procedures and quality control including documentation and stakeholder communication
- Troubleshoot incidents following ITIL structure and incident management lifecycle
- Test system integrity application designs and infrastructure processes implementing improvements as needed
- Oversee release management and deploy applications in production environments in line with enterprise standards
- Support production releases or urgent changes during weekends or non-business hours as needed
Key Responsibilities
- Ensure high availability and performance of Big Data infrastructure
- Maintain secure resilient and scalable cloud environments
- Drive automation and continuous improvement across infrastructure and deployment pipelines
- Own stakeholder communication to ensure transparency and alignment during release cycles and critical incidents
Mandatory Skills and Qualifications:
- Bachelors degree in Computer Science Information Systems or related IT disciplines
- Exceptional Python programming and scripting skills with hands-on experience using Python for automation infrastructure and data-related tasks.
- Experience with major Cloud Service Providers (AWS Huawei).
- Expertise in cloud-based system support: provisioning automation monitoring security resilience backup and recovery.
- Familiarity with cloud-native infrastructure (e.g. Serverless Kubernetes).
- Proficiency in DevOps and SRE tools (Terraform Ansible Git).
- Background in mission-critical high-performance large-scale systems.
- Strong analytical problem-solving interpersonal and presentation skills.
- Fluent in written and spoken English and Chinese.
Preferred Skills
- Experience with observability and monitoring tools such as Grafana Kibana Prometheus or equivalent
- Exposure to advanced alerting and dashboarding for system health and performance
- Familiarity with log aggregation and distributed tracing tools (e.g. ELK stack)
Company Introduction:
ITD SZ
港交所科技深圳有限公司是2016年12月28日于深圳市前海自贸区成立的外商独资企业
作为港交所的技术子公司港交所科技深圳有限公司主要是为集团及其附属公司提供计算机软件计算机硬件信息系统云存储云计算物联网和计算机网络的开发技术服务技术咨询技术转让经济信息咨询企业管理咨询商务信息咨询商业信息咨询信息系统设计集成运行维护数据库管理大数据分析以承接服务外包方式提供系统应用管理和维护信息技术支持管理数据处理等信息技术和业务流程外包服务
Required Experience:
Chief
Location:CN-Shenzhen-HyQShift:Standard - 40 Hours (China)Scheduled Weekly Hours:40Worker Type:PermanentJob Summary:The AVP under Site Reliability Engineering (SRE) in the Chief Data Office plays a critical role in maintaining and enhancing the reliability performance and scalability of our enterpris...
Location:
CN-Shenzhen-HyQ
Shift:
Standard - 40 Hours (China)
Scheduled Weekly Hours:
40
Worker Type:
Permanent
Job Summary:
The AVP under Site Reliability Engineering (SRE) in the Chief Data Office plays a critical role in maintaining and enhancing the reliability performance and scalability of our enterprise data platform. This position ensures the stability of production testing and development environments that support Big Data and Analytics initiatives. You will leverage expertise in DevOps practices Infrastructure as Code and Kubernetes Architecture to drive automation streamline operations and proactively resolve system issues. A strong background in managing cloud-based large-scale data infrastructure is essential along with advanced Python scripting skills for automation and infrastructure management.
Job Duties:
Job Duties
- Build and maintain cloud-based Big Data and Analytics Platforms including:
- Enterprise Data Lake
- Data Governance and Management Platforms
- Self-service BI and Augmented Analytics Platforms
- Proactively managing production services and data pipelines to ensure availability and system health
- Improve system reliability through engineering solutions and advanced tooling for monitoring automation and fault tolerance
- Develop and maintain CI/CD tools microservices and solutions to meet customer service standards and minimize errors
- Manage release procedures and quality control including documentation and stakeholder communication
- Troubleshoot incidents following ITIL structure and incident management lifecycle
- Test system integrity application designs and infrastructure processes implementing improvements as needed
- Oversee release management and deploy applications in production environments in line with enterprise standards
- Support production releases or urgent changes during weekends or non-business hours as needed
Key Responsibilities
- Ensure high availability and performance of Big Data infrastructure
- Maintain secure resilient and scalable cloud environments
- Drive automation and continuous improvement across infrastructure and deployment pipelines
- Own stakeholder communication to ensure transparency and alignment during release cycles and critical incidents
Mandatory Skills and Qualifications:
- Bachelors degree in Computer Science Information Systems or related IT disciplines
- Exceptional Python programming and scripting skills with hands-on experience using Python for automation infrastructure and data-related tasks.
- Experience with major Cloud Service Providers (AWS Huawei).
- Expertise in cloud-based system support: provisioning automation monitoring security resilience backup and recovery.
- Familiarity with cloud-native infrastructure (e.g. Serverless Kubernetes).
- Proficiency in DevOps and SRE tools (Terraform Ansible Git).
- Background in mission-critical high-performance large-scale systems.
- Strong analytical problem-solving interpersonal and presentation skills.
- Fluent in written and spoken English and Chinese.
Preferred Skills
- Experience with observability and monitoring tools such as Grafana Kibana Prometheus or equivalent
- Exposure to advanced alerting and dashboarding for system health and performance
- Familiarity with log aggregation and distributed tracing tools (e.g. ELK stack)
Company Introduction:
ITD SZ
港交所科技深圳有限公司是2016年12月28日于深圳市前海自贸区成立的外商独资企业
作为港交所的技术子公司港交所科技深圳有限公司主要是为集团及其附属公司提供计算机软件计算机硬件信息系统云存储云计算物联网和计算机网络的开发技术服务技术咨询技术转让经济信息咨询企业管理咨询商务信息咨询商业信息咨询信息系统设计集成运行维护数据库管理大数据分析以承接服务外包方式提供系统应用管理和维护信息技术支持管理数据处理等信息技术和业务流程外包服务
Required Experience:
Chief
View more
View less