Employer Active
Job Alert
You will be updated with latest job alerts via emailJob Alert
You will be updated with latest job alerts via emailAt Medtronic you can begin a life-long career of exploration and innovation while helping champion healthcare access and equity for all. Youll lead with purpose breaking down barriers to innovation in a more connected compassionate world.
Responsibilities may include the following and other duties may be assigned:
Design develop and maintain scalable and efficient ETL pipelines using PySpark.
Work with structured and unstructured data from various sources.
Optimize and tune PySpark applications for performance and scalability.
Collaborate with data scientists and analysts to understand data requirements review Business Requirement documents and deliver high-quality datasets.
Implement data quality checks and ensure data integrity.
Monitor and troubleshoot data pipeline issues and ensure timely resolution.
Document technical specifications and maintain comprehensive documentation for data pipelines.
Stay up to date with the latest trends and technologies in big data and distributed computing.
Required Knowledge and Experience:
Bachelors degree in computer science Engineering or a related field.
4-5 years of experience in data engineering with a focus on PySpark.
Proficiency in Python and Spark with strong coding and debugging skills.
Strong knowledge of SQL and experience with relational databases (e.g. PostgreSQL MySQL SQL Server).
Hands-on experience with cloud platforms such as AWS Azure or Google Cloud Platform (GCP).
Experience with data warehousing solutions like Redshift Snowflake Databricks or Google BigQuery.
Familiarity with data lake architectures and data storage solutions.
Experience with big data technologies such as Hadoop Hive and Kafka.
Excellent problem-solving skills and the ability to troubleshoot complex issues.
Strong communication and collaboration skills with the ability to work effectively in a team environment.
Preferred Skills:
Experience with Databricks.
Experience with orchestration tools like Apache Airflow or AWS Step Functions.
Knowledge of machine learning workflows and experience working with data scientists.
Understanding of data security and governance best practices.
Familiarity with streaming data platforms and real-time data processing.
Knowledge of CI/CD pipelines and version control systems (e.g. Git).
Physical Job Requirements
The above statements are intended to describe the general nature and level of work being performed by employees assigned to this position but they are not an exhaustive list of all the required responsibilities and skills of this position.
Medtronic offers a competitive Salary and flexible Benefits Package
A commitment to our employees lives at the core of our values. We recognize their contributions. They share in the success they help to create. We offer a wide range of benefits resources and competitive compensation plans designed to support you at every career and life stage.
We lead global healthcare technology and boldly attack the most challenging health problems facing humanity by searching out and finding solutions.
Our Mission to alleviate pain restore health and extend life unites a global team of 95000 passionate people.
We are engineers at heart putting ambitious ideas to work to generate real solutions for real people. From the R&D lab to the factory floor to the conference room every one of us experiments creates builds improves and solves. We have the talent diverse perspectives and guts to engineer the extraordinary.
Learn more about our business mission and our commitment to diversity here
Full-Time