ETL/ELT Process Design & Implementation:
- Design and implement robust ETL/ELT processes to extract transform and load data from diverse sources into data warehouses and lakes.
- Ensure efficient data pipeline architectures that support data quality integrity and consistency.
Data Pipeline Management:
- Develop and optimize data pipelines for performance and reliability managing structured and unstructured data across databases (Oracle MySQL MongoDB DB2).
- Monitor and troubleshoot pipeline issues implementing solutions to maintain stability.
Data Preprocessing & Quality Assurance:
- Oversee data preprocessing cleaning and validation processes to meet quality standards.
- Collaborate with crossfunctional teams to understand data requirements and deliver effective solutions.
Data Analysis & Visualization:
- Utilize Jupyter Notebook and Power BI to analyze data and create insightful visualizations.
- Document data workflows processes and methodologies for clarity and continuity.
Team Leadership & Collaboration:
- Build and lead a data engineering team fostering a culture of collaboration and innovation.
- Promote knowledge sharing and align data initiatives with organizational goals.
Requirements
Bachelor s degree in Computer Science Information Technology or a related field.
12 years of handson data engineering experience with expertise in ETL/ELT processes and data quality assurance.
Proficient in Python PySpark and Scala with experience in VS Code and data pipeline management.
Strong knowledge of databases (Oracle MySQL MongoDB DB2) and data visualization tools (Power BI Jupyter Notebook).
Experience with cloud platforms (AWS Azure) and an understanding of data warehousing best practices.
Knowledge of big data technologies and frameworks is a plus.
Excellent problemsolving skills and a proactive mindset.
Bachelor s degree in Computer Science, Information Technology, or a related field. 12+ years of hands-on data engineering experience, with expertise in ETL/ELT processes and data quality assurance. Proficient in Python, PySpark, and Scala, with experience in VS Code and data pipeline management. Strong knowledge of databases (Oracle, MySQL, MongoDB, DB2) and data visualization tools (Power BI, Jupyter Notebook). Experience with cloud platforms (AWS, Azure) and an understanding of data warehousing best practices. Knowledge of big data technologies and frameworks is a plus. Excellent problem-solving skills and a proactive mindset.