Roles andResponsibilities:
- Responsible for developing and maintaining applications with PySpark
- Contribute to the overall design and architecture of the application developed and deployed.
- Performance Tuning wrt to executor sizing and other environmental parameters code optimization partitions tuning etc
- Interact with business users to understand requirements and troubleshoot issues.
- Implement Projects based on functional specifications.
MustHaveSkills:
- Relevant Experience: 36 Years
- SQL Mandatory
- Python Mandatory
- SparkSQL Mandatory
- PySpark Mandatory
- Hive Mandatory
- HDFS and Spark Mandatory
- Scala Advantage
- Apache Airflow Advantage
Requirements
36 Years of Experience
Must Have: PySpark/Spark Python SQL Knowledge on Hadoop ecosystem
Good to have: Airflow Scala