Databrick Pyspark Engineer

Virtusa

Not Interested
Bookmark
Report This Job

profile Job Location:

Chennai - India

profile Monthly Salary: Not Disclosed
Posted on: 03-09-2025
Vacancies: 1 Vacancy

Job Summary

Key Responsibilities:Design develop and maintain scalable data pipelines using Apache Spark on Databricks. Write efficient and production-ready PySpark or Scala code for data transformation and ETL processes. Integrate data from various structured and unstructured sources into a unified platform. Implement Delta Lake and manage data versioning updates and schema evolution. Optimize data processing workflows for performance scalability and cost efficiency. Collaborate with data scientists analysts and business stakeholders to deliver high-quality datasets. Implement data quality checks validation routines and logging mechanisms. Monitor and debug production jobs using Databricks jobs notebooks and clusters. Ensure security privacy and compliance standards are met throughout the data lifecycle. Provide guidance and mentorship to junior team Skills & Qualifications:4 to 6 years of experience in Big Data development. Hands-on experience with Databricks (including Workflows Notebooks Delta Live Tables Unity Catalog). Strong programming skills in PySpark and/or Scala. Solid understanding of Delta Lake architecture. Proficient in SQL for data analysis and transformation. Experience with cloud platforms such as Azure (Azure Data Lake Data Factory Synapse) or AWS (S3 Glue Redshift). Familiarity with CI/CD for Databricks deployments (e.g. using GitHub Actions Azure DevOps). Knowledge of data governance cataloguing and security best practices. Experience working in an Agile/Scrum Skills:Experience with Databricks Unity Catalog and Delta Live Tables. Exposure to machine learning workflows in Databricks. Experience with Apache Airflow Kafka or other orchestration/messaging tools. Certifications such as Databricks Certified Data Engineer Associate/Professional Azure or AWS certification.
Key Responsibilities:Design develop and maintain scalable data pipelines using Apache Spark on Databricks. Write efficient and production-ready PySpark or Scala code for data transformation and ETL processes. Integrate data from various structured and unstructured sources into a unified platform. Im...
View more view more

Key Skills

  • ASP.NET
  • Health Education
  • Fashion Designing
  • Fiber
  • Investigation

About Company

Inside every Virtusan is a spirit defined by the drive to explore new frontiers, an intellectual curiosity, a need to challenge the status quo, and the inspiration to mind the greater good–all while impacting the bottom line. It all adds up to a culture of innovation.

View Profile View Profile