Work Schedule
Other
Environmental Conditions
Office
Job Description
Summarized Purpose:
We are offering an outstanding opportunity to join Thermo Fisher Scientific as a Data Warehouse this role you will focus on designing building and optimizing database-centric solutions that power our analytical and operational data platforms. You will play a key role in developing robust scalable data pipelines and warehouse structures -primarily on AWS Redshiftsupporting data-intensive workloads and downstream reporting applications.
Education/Experience:
- Bachelors degree or comparable experience in Computer Science Information Science or a related area of study
- Around 3 years of experience in database development data engineering or a related field
- Equivalent combinations of education training and experience will also be considered
Major Job Responsibilities:
- Design implement and refine data warehouse schemas (e.g. star/snowflake models) and data pipelines in AWS Redshift or similar RDBMS platforms.
- Build and manage SQL-centric ETL/ELT procedures for the processing transformation and merging of data from diverse origins.
- Improve database performance through query optimization and efficient data processing.
- Collaborate with stakeholders to translate data needs into scalable and maintainable data structures.
- Support data quality validation and reconciliation processes to ensure accurate and reliable datasets.
- Engage with AWS services including Lambda Step Functions and S3 for orchestrating and automating data workflows.
- Participate in design reviews documentation and testing activities to ensure adherence to quality and compliance standards.
- Collaborate with Operations and DevOps teams to deploy and monitor data workflows using CI/CD pipelines where applicable.
- Troubleshoot production issues analyze root causes and propose balanced solutions.
- Leverage AI-assisted development tools to improve query efficiency code refactoring and documentation.
Knowledge Skills and Abilities:
- Strong hands-on SQL development skills including complex queries window functions joins and analytical operations.
- Mastery in data modeling and grasp of data warehousing principles (ETL/ELT dimensional modeling slowly altering dimensions).
- Experience working with large relational databases (e.g. Redshift PostgreSQL SQL Server MySQL Oracle).
- Knowledge of AWS cloud services especially S3 Lambda Redshift and Step Functions.
- Familiarity with Python or NodeJS for scripting automation or data workflows based on Lambda.
- Excellent analytical and problem-solving skills with attention to detail.
- Strong communication and collaboration skills in a team-oriented environment.
Must Have skills:
- Advanced SQL and RDBMS experience Ability to develop and optimize queries stored procedures and data transformations for large-scale data workloads.
- Data warehousing and ETL/ELT Practical experience designing and maintaining data warehouse environments and data pipelines.
- Practical exposure to AWS services Hands-on experience with AWS data services like Redshift S3 Lambda and Step Functions.
- Skilled in Python or NodeJS programming Capable of using a programming language for automation or data integration purposes.
- Data modeling and schema creation Experience crafting normalized and dimensional schemas for analytics and reporting.
Good to have skills:
- Exposure to data Lakehouse or big data environments (Databricks Snowflake or similar).
- Knowledge of AI-assisted or modern query optimization tools and practices.
Working Hours:
India: 05:30 PM to 02:30 AM IST
Philippines: 08:00 PM to 05:00 AM PST
Work ScheduleOtherEnvironmental ConditionsOfficeJob DescriptionSummarized Purpose:We are offering an outstanding opportunity to join Thermo Fisher Scientific as a Data Warehouse this role you will focus on designing building and optimizing database-centric solutions that power our analytical and op...
Work Schedule
Other
Environmental Conditions
Office
Job Description
Summarized Purpose:
We are offering an outstanding opportunity to join Thermo Fisher Scientific as a Data Warehouse this role you will focus on designing building and optimizing database-centric solutions that power our analytical and operational data platforms. You will play a key role in developing robust scalable data pipelines and warehouse structures -primarily on AWS Redshiftsupporting data-intensive workloads and downstream reporting applications.
Education/Experience:
- Bachelors degree or comparable experience in Computer Science Information Science or a related area of study
- Around 3 years of experience in database development data engineering or a related field
- Equivalent combinations of education training and experience will also be considered
Major Job Responsibilities:
- Design implement and refine data warehouse schemas (e.g. star/snowflake models) and data pipelines in AWS Redshift or similar RDBMS platforms.
- Build and manage SQL-centric ETL/ELT procedures for the processing transformation and merging of data from diverse origins.
- Improve database performance through query optimization and efficient data processing.
- Collaborate with stakeholders to translate data needs into scalable and maintainable data structures.
- Support data quality validation and reconciliation processes to ensure accurate and reliable datasets.
- Engage with AWS services including Lambda Step Functions and S3 for orchestrating and automating data workflows.
- Participate in design reviews documentation and testing activities to ensure adherence to quality and compliance standards.
- Collaborate with Operations and DevOps teams to deploy and monitor data workflows using CI/CD pipelines where applicable.
- Troubleshoot production issues analyze root causes and propose balanced solutions.
- Leverage AI-assisted development tools to improve query efficiency code refactoring and documentation.
Knowledge Skills and Abilities:
- Strong hands-on SQL development skills including complex queries window functions joins and analytical operations.
- Mastery in data modeling and grasp of data warehousing principles (ETL/ELT dimensional modeling slowly altering dimensions).
- Experience working with large relational databases (e.g. Redshift PostgreSQL SQL Server MySQL Oracle).
- Knowledge of AWS cloud services especially S3 Lambda Redshift and Step Functions.
- Familiarity with Python or NodeJS for scripting automation or data workflows based on Lambda.
- Excellent analytical and problem-solving skills with attention to detail.
- Strong communication and collaboration skills in a team-oriented environment.
Must Have skills:
- Advanced SQL and RDBMS experience Ability to develop and optimize queries stored procedures and data transformations for large-scale data workloads.
- Data warehousing and ETL/ELT Practical experience designing and maintaining data warehouse environments and data pipelines.
- Practical exposure to AWS services Hands-on experience with AWS data services like Redshift S3 Lambda and Step Functions.
- Skilled in Python or NodeJS programming Capable of using a programming language for automation or data integration purposes.
- Data modeling and schema creation Experience crafting normalized and dimensional schemas for analytics and reporting.
Good to have skills:
- Exposure to data Lakehouse or big data environments (Databricks Snowflake or similar).
- Knowledge of AI-assisted or modern query optimization tools and practices.
Working Hours:
India: 05:30 PM to 02:30 AM IST
Philippines: 08:00 PM to 05:00 AM PST
View more
View less