As part of our team you will enhance thousands of compute and big data pipelines to deliver greater scalability reliability and efficiency. By leveraging innovative approaches with machine learning and large language models you will improve pipeline quality optimize Spark and Kubernetes resource utilization and create automation that accelerates developer agility.
- Bachelors degree in Computer Science Computer Engineering or a related field.
- 3 years of experience with large-scale data processing and pipelines.
- Proficiency in Scala Python and scripting languages.
- Experience in and solid understanding of distributed systems performance tuning and resource optimization.
- Strong hands-on expertise with Apache Spark and the Hadoop ecosystem.
- Experience developing or applying machine learning techniques or LLM-based agentic workflows for data pipeline optimization and data quality improvements.
- Knowledge of cost optimization strategies for big data infrastructure.
As part of our team you will enhance thousands of compute and big data pipelines to deliver greater scalability reliability and efficiency. By leveraging innovative approaches with machine learning and large language models you will improve pipeline quality optimize Spark and Kubernetes resource uti...
As part of our team you will enhance thousands of compute and big data pipelines to deliver greater scalability reliability and efficiency. By leveraging innovative approaches with machine learning and large language models you will improve pipeline quality optimize Spark and Kubernetes resource utilization and create automation that accelerates developer agility.
- Bachelors degree in Computer Science Computer Engineering or a related field.
- 3 years of experience with large-scale data processing and pipelines.
- Proficiency in Scala Python and scripting languages.
- Experience in and solid understanding of distributed systems performance tuning and resource optimization.
- Strong hands-on expertise with Apache Spark and the Hadoop ecosystem.
- Experience developing or applying machine learning techniques or LLM-based agentic workflows for data pipeline optimization and data quality improvements.
- Knowledge of cost optimization strategies for big data infrastructure.
View more
View less