Location:Durham NC(Hybrid)
Mandatory: Experience in Clinical Data is a plus.
Python SQL Snowflake DBT Core Airflow vsCode are mandatory.
We are seeking a highly motivated and skilled Data Engineer to join our data and analytics team. The ideal candidate will be responsible for designing building and maintaining robust and scalable data pipelines and infrastructure using modern cloud-based technologies. The role requires a hands-on expert in Snowflake DBT Airflow and Python to manage complex data transformations and orchestrate workflows. Clinical data experience is strongly preferred. This position requires strong technical proficiency excellent problem-solving abilities and a high degree of ownership to ensure the timely delivery of high-quality data products.
Key Responsibilities - Data Pipeline Development: Design build and maintain scalable and reliable data pipelines using Airflow to ingest transform and load data into Snowflake.
- Data Modeling and Transformation: Develop and manage ELT/ETL processes and data models using DBT Core to produce clean validated and analytics-ready datasets in Snowflake.
- Snowflake Architecture: Design and optimize the Snowflake environment including data warehouses performance tuning and access control for data integrity and security.
- Python Scripting and AI Tools: Utilize Python for data scripting automation API integrations and to support Python-based AI and machine learning tools within the data pipeline.
- SQL Expertise: Write complex and highly performant SQL queries to define data transformations conduct data analysis and troubleshoot issues.
- Data Quality and Validation: Develop and implement rigorous data validation monitoring and testing frameworks to ensure the accuracy completeness and integrity of all data.
- Debugging and Troubleshooting: Proactively identify debug and resolve issues related to data pipelines performance bottlenecks and infrastructure in a timely manner.
- Clinical Data Management (Preferred): Work with clinical and healthcare data ensuring compliance with relevant data privacy and security regulations such as HIPAA.
- Collaboration and Communication: Work closely with cross-functional teams including data scientists analysts and business stakeholders to understand requirements and deliver data-driven solutions effectively.
- Ownership and Accountability: Take full ownership of projects from inception to completion demonstrating a proactive approach and a commitment to delivering high-quality work on schedule.
Role:Data Engineer Location:Durham NC(Hybrid) Mandatory: Experience in Clinical Data is a plus. Python SQL Snowflake DBT Core Airflow vsCode are mandatory. We are seeking a highly motivated and skilled Data Engineer to join our data and analytics team. The ideal candidate will be responsible fo...
Location:Durham NC(Hybrid)
Mandatory: Experience in Clinical Data is a plus.
Python SQL Snowflake DBT Core Airflow vsCode are mandatory.
We are seeking a highly motivated and skilled Data Engineer to join our data and analytics team. The ideal candidate will be responsible for designing building and maintaining robust and scalable data pipelines and infrastructure using modern cloud-based technologies. The role requires a hands-on expert in Snowflake DBT Airflow and Python to manage complex data transformations and orchestrate workflows. Clinical data experience is strongly preferred. This position requires strong technical proficiency excellent problem-solving abilities and a high degree of ownership to ensure the timely delivery of high-quality data products.
Key Responsibilities - Data Pipeline Development: Design build and maintain scalable and reliable data pipelines using Airflow to ingest transform and load data into Snowflake.
- Data Modeling and Transformation: Develop and manage ELT/ETL processes and data models using DBT Core to produce clean validated and analytics-ready datasets in Snowflake.
- Snowflake Architecture: Design and optimize the Snowflake environment including data warehouses performance tuning and access control for data integrity and security.
- Python Scripting and AI Tools: Utilize Python for data scripting automation API integrations and to support Python-based AI and machine learning tools within the data pipeline.
- SQL Expertise: Write complex and highly performant SQL queries to define data transformations conduct data analysis and troubleshoot issues.
- Data Quality and Validation: Develop and implement rigorous data validation monitoring and testing frameworks to ensure the accuracy completeness and integrity of all data.
- Debugging and Troubleshooting: Proactively identify debug and resolve issues related to data pipelines performance bottlenecks and infrastructure in a timely manner.
- Clinical Data Management (Preferred): Work with clinical and healthcare data ensuring compliance with relevant data privacy and security regulations such as HIPAA.
- Collaboration and Communication: Work closely with cross-functional teams including data scientists analysts and business stakeholders to understand requirements and deliver data-driven solutions effectively.
- Ownership and Accountability: Take full ownership of projects from inception to completion demonstrating a proactive approach and a commitment to delivering high-quality work on schedule.
View more
View less