USC/GC - comfortable to work in Eastern & Central time
Position: PySpark/Java Developer
Location: Remote
Job Description:
We are seeking an experienced PySpark/Java Developer to join a leading healthcare analytics team. The ideal candidate will design and develop scalable data solutions build and maintain ETL pipelines and work with large datasets to support data processing validation and reporting. This role involves collaborating with cross-functional teams troubleshooting performance issues and ensuring high-quality maintainable code.
Key Responsibilities:
- Develop and maintain ETL pipelines and data processing applications
- Work with large-scale structured and unstructured data using Spark
- Optimize SQL queries and resolve performance issues
- Design test and deploy scalable data solutions
- Collaborate on requirements design and implementation
- Support data validation reporting and production issues
Required Skills:
- Strong experience with PySpark Java and Python
- Hands-on experience with Hadoop ecosystem (Hive Spark Impala Kafka Sqoop Oozie Yarn)
- Solid ETL and data engineering experience
- Expertise in SQL Server and performance tuning
- Experience handling large data volumes
- Healthcare data/claims experience preferred
USC/GC - comfortable to work in Eastern & Central time Position: PySpark/Java Developer Location: Remote Job Description: We are seeking an experienced PySpark/Java Developer to join a leading healthcare analytics team. The ideal candidate will design and develop scalable data solutions build and m...
USC/GC - comfortable to work in Eastern & Central time
Position: PySpark/Java Developer
Location: Remote
Job Description:
We are seeking an experienced PySpark/Java Developer to join a leading healthcare analytics team. The ideal candidate will design and develop scalable data solutions build and maintain ETL pipelines and work with large datasets to support data processing validation and reporting. This role involves collaborating with cross-functional teams troubleshooting performance issues and ensuring high-quality maintainable code.
Key Responsibilities:
- Develop and maintain ETL pipelines and data processing applications
- Work with large-scale structured and unstructured data using Spark
- Optimize SQL queries and resolve performance issues
- Design test and deploy scalable data solutions
- Collaborate on requirements design and implementation
- Support data validation reporting and production issues
Required Skills:
- Strong experience with PySpark Java and Python
- Hands-on experience with Hadoop ecosystem (Hive Spark Impala Kafka Sqoop Oozie Yarn)
- Solid ETL and data engineering experience
- Expertise in SQL Server and performance tuning
- Experience handling large data volumes
- Healthcare data/claims experience preferred
View more
View less