Job Description
We are looking for a highly skilled Senior Data Engineer with strong experience in Java Apache Spark PySpark and AWS Glue to design build and maintain scalable data processing solutions on AWS. The ideal candidate should have hands-on experience in cloud-native data engineering and modern DevOps practices.
Key Responsibilities
- Design develop and optimize scalable data pipelines using Apache Spark (Java & PySpark).
- Build and manage ETL workflows using AWS Glue.
- Write high-quality efficient and reusable Core Java code.
- Develop and maintain cloud infrastructure using Terraform (IaC).
- Deploy and manage applications in containerized environments using Kubernetes.
- Work closely with data architects analysts and stakeholders to deliver reliable data solutions.
- Ensure performance tuning monitoring and troubleshooting of data pipelines.
- Follow best practices for security scalability and reliability on AWS.
Required Skills
- Strong hands-on experience with Core Java and Java-based Spark applications.
- Solid experience in PySpark and distributed data processing.
- Proven expertise in AWS Glue (Jobs Crawlers Data Catalog).
- Good understanding of AWS services (S3 IAM CloudWatch etc.).
- Experience with Terraform for infrastructure provisioning.
- Hands-on experience with Kubernetes and container orchestration.
- Strong coding debugging and performance optimization skills.
Nice to Have
- Experience with CI/CD pipelines
- Exposure to data lakes and big data architectures
- AWS certifications
Job Description We are looking for a highly skilled Senior Data Engineer with strong experience in Java Apache Spark PySpark and AWS Glue to design build and maintain scalable data processing solutions on AWS. The ideal candidate should have hands-on experience in cloud-native data engineering and m...
Job Description
We are looking for a highly skilled Senior Data Engineer with strong experience in Java Apache Spark PySpark and AWS Glue to design build and maintain scalable data processing solutions on AWS. The ideal candidate should have hands-on experience in cloud-native data engineering and modern DevOps practices.
Key Responsibilities
- Design develop and optimize scalable data pipelines using Apache Spark (Java & PySpark).
- Build and manage ETL workflows using AWS Glue.
- Write high-quality efficient and reusable Core Java code.
- Develop and maintain cloud infrastructure using Terraform (IaC).
- Deploy and manage applications in containerized environments using Kubernetes.
- Work closely with data architects analysts and stakeholders to deliver reliable data solutions.
- Ensure performance tuning monitoring and troubleshooting of data pipelines.
- Follow best practices for security scalability and reliability on AWS.
Required Skills
- Strong hands-on experience with Core Java and Java-based Spark applications.
- Solid experience in PySpark and distributed data processing.
- Proven expertise in AWS Glue (Jobs Crawlers Data Catalog).
- Good understanding of AWS services (S3 IAM CloudWatch etc.).
- Experience with Terraform for infrastructure provisioning.
- Hands-on experience with Kubernetes and container orchestration.
- Strong coding debugging and performance optimization skills.
Nice to Have
- Experience with CI/CD pipelines
- Exposure to data lakes and big data architectures
- AWS certifications
View more
View less