Employer Active
Job Alert
You will be updated with latest job alerts via emailJob Alert
You will be updated with latest job alerts via emailLocation : Mumbai
Notice Period : 15 Days
Experience Required : 1 years
Mandatory Skillset : Pyspark/Python
Job description:
Build and Operate Data Pipelines: Develop and maintain robust and scalable data pipelines to ensure the efficient and reliable flow of data within the organization.
ETL Optimization: Design code and finetune big data processes using Apache Spark for optimal performance and data processing efficiency.
Proficiency in SQL: Utilize SQL to extract transform and analyze data from various sources ensuring data accuracy and integrity.
Hadoop Architecture Knowledge: Demonstrate a solid understanding of Hadoop architecture and its components to support our big data infrastructure.
Proficiency in Hive: Work with Hive for data warehousing and querying creating structured datasets for analytical purposes.
Experience in Workflow Orchestration: Collaborate with teams to design and implement workflow orchestrations using tools such as Informatics power center.
Full Time