Role: Data Engineer (SQL PySpark and Python)
Location: Pittsburgh PA Hybrid ( 3 to 4 days from office)
6-12 Months Contract Possible Extension
- Design develop and optimize data pipelines and big data applications using SQL PySpark and Python.
Key Responsibilities:
- This role is pivotal in enabling organizations to process analyze and manage large datasets efficiently on modern enterprise data platforms.
- Develop and maintain complex SQL queries and stored procedures for data extraction transformation and loading (ETL).
- Build and optimize scalable data pipelines and data processing workflows using PySpark and Python.
- Collaborate with data engineers scientists and analysts to understand and fulfill data requirements.
- Ensure data quality integrity and consistent performance across big data environments.
- Debug monitor and fine-tune data jobs for optimal performance.
- Document code and processes; adhere to best practices for coding and performance.