Job Summary:
The Data Scientist will participate in product teams to design develop and maintain cloud-based data and analytics products with a focus on data lake and lakehouse structures. This role involves migrating legacy data pipelines supporting the development of standards and participating in code reviews to ensure code quality and promote knowledge sharing.
Location: Toronto Ontario Canada
Responsibilities:
- Participate in product teams to analyze systems requirements architect design code and implement cloud-based data and analytics products that conform to standards.
- Design create and maintain cloud-based data lake and lakehouse structures automated data pipelines and analytics models.
- Liaise with IT colleagues to implement products conduct reviews resolve operational problems and support business partners in effective use of cloud-based data and analytics products.
- Analyze complex technical issues identify alternatives and recommend solutions.
- Support the migration of legacy data pipelines to modernized Databricks-based solutions leveraging Delta Lake and native orchestration capabilities.
- Support the development of standards and a reusable framework that streamlines pipeline creation.
- Participate in code reviews and prepare/conduct knowledge transfer to maintain code quality promote team knowledge sharing and enforce development standards across collaborative data projects.
Required Skills & Certifications:
- 5 years of experience in an Azure environment.
- 5 years of experience in Data Engineering with ADF and Databricks.
- 5 years of programming experience with Python and SQL.
- Experience in multiple cloud-based data and analytics platforms and coding/programming/scripting tools.
- Experience with designing creating and maintaining cloud-based data lake and lakehouse structures automated data pipelines analytics models.
- Strong background in building and orchestrating data pipelines using services like Azure Data Factory and Databricks.
- Demonstrated ability to organize and manage data in a lakehouse following medallion architecture.
- Proficient in using Python and SQL for data engineering and analytics development.
- Familiar with CI/CD practices and tools for automating deployment of data solutions and managing code lifecycle.
- Comfortable conducting and participating in peer code reviews in GitHub.
- Experience in assessing client information technology needs and objectives.
- Experience in problem-solving to resolve complex multi-component failures.
- Experience in preparing knowledge transfer documentation and conducting knowledge transfer.
- Experience working on an Agile team.
Preferred Skills & Certifications:
Special Considerations:
Scheduling: