Overview
Susquehanna is looking for an experienced ETL Developer to join our technology team at our headquarters in Bala Cynwyd PA just outside of Center City Philadelphia. As a member of our ETL team you will participate the design and development of our ETL infrastructure. You will work directly with technical and business teams to analyze their data needs and develop solutions for both batch and continuous data movements using enterprise class tools and custom scripting methods.
This role will give you exposure to cutting-edge distributed database environments that support critical business activities. We are looking for someone who is excited to join our efforts as we evolve evaluate and promote newer technologies within this space across our firm globally.
In this role you will:
- Design develop and optimize PySpark data processing workflows for large-scale datasets.
- Build and maintain real-time and batch data pipelines leveraging Apache Kafka.
- Write clean efficient and maintainable Python code for data transformation ETL and automation.
- Develop shell scripts and other automation scripts to support data workflows and operational tasks.
- Work with relational databases (e.g. PostgreSQL MySQL SQL Server Oracle) to write efficient SQL queries manage schemas and optimize performance.
- Collaborate closely with data engineers analysts and platform teams to deliver endtoend data solutions.
- Troubleshoot and improve existing pipelines ensuring performance reliability and scalability.
- Follow best practices in version control CI/CD documentation and code reviews.
What were looking for
- Bachelors degree in Computer Science or a related technical discipline is required
- 5 years of handson development experience in data engineering or backend engineering roles.
- Strong proficiency in:
- PySpark (RDDs DataFrames Spark SQL performance tuning)
- Apache Kafka (producers/consumers topics partitions schema management)
- Python (data manipulation modular code error handling)
- Scripting (Bash Shell or similar)
- Experience working with relational databases and writing optimized SQL queries.
- Solid understanding of distributed systems and largescale data processing concepts.
- Familiarity with ETL best practices data modeling and pipeline orchestration
- Basic understanding of LLM operations (LLMOps) including prompt logging model monitoring and evaluation
- Knowledge of Spark cluster management resource optimization and tuning strategies is preferred.
- Experience with workflow orchestration tools such as Apache Airflow is a plus
SIG does not accept unsolicited resumes from recruiters or search firms. Any resume or referral submitted in the absence of a signed agreement will become the property of SIG and no fee will be paid.
Required Experience:
Senior IC
OverviewSusquehanna is looking for an experienced ETL Developer to join our technology team at our headquarters in Bala Cynwyd PA just outside of Center City Philadelphia. As a member of our ETL team you will participate the design and development of our ETL infrastructure. You will work directly wi...
Overview
Susquehanna is looking for an experienced ETL Developer to join our technology team at our headquarters in Bala Cynwyd PA just outside of Center City Philadelphia. As a member of our ETL team you will participate the design and development of our ETL infrastructure. You will work directly with technical and business teams to analyze their data needs and develop solutions for both batch and continuous data movements using enterprise class tools and custom scripting methods.
This role will give you exposure to cutting-edge distributed database environments that support critical business activities. We are looking for someone who is excited to join our efforts as we evolve evaluate and promote newer technologies within this space across our firm globally.
In this role you will:
- Design develop and optimize PySpark data processing workflows for large-scale datasets.
- Build and maintain real-time and batch data pipelines leveraging Apache Kafka.
- Write clean efficient and maintainable Python code for data transformation ETL and automation.
- Develop shell scripts and other automation scripts to support data workflows and operational tasks.
- Work with relational databases (e.g. PostgreSQL MySQL SQL Server Oracle) to write efficient SQL queries manage schemas and optimize performance.
- Collaborate closely with data engineers analysts and platform teams to deliver endtoend data solutions.
- Troubleshoot and improve existing pipelines ensuring performance reliability and scalability.
- Follow best practices in version control CI/CD documentation and code reviews.
What were looking for
- Bachelors degree in Computer Science or a related technical discipline is required
- 5 years of handson development experience in data engineering or backend engineering roles.
- Strong proficiency in:
- PySpark (RDDs DataFrames Spark SQL performance tuning)
- Apache Kafka (producers/consumers topics partitions schema management)
- Python (data manipulation modular code error handling)
- Scripting (Bash Shell or similar)
- Experience working with relational databases and writing optimized SQL queries.
- Solid understanding of distributed systems and largescale data processing concepts.
- Familiarity with ETL best practices data modeling and pipeline orchestration
- Basic understanding of LLM operations (LLMOps) including prompt logging model monitoring and evaluation
- Knowledge of Spark cluster management resource optimization and tuning strategies is preferred.
- Experience with workflow orchestration tools such as Apache Airflow is a plus
SIG does not accept unsolicited resumes from recruiters or search firms. Any resume or referral submitted in the absence of a signed agreement will become the property of SIG and no fee will be paid.
Required Experience:
Senior IC
View more
View less