Responsibilities:Design develop and optimize data pipelines and workflows for big data components application interfaces and solution enablers while ensuring principal architecture integrity is and maintain distributed systems using Hadoop Spark or similar large datasets to extract meaningful insights and data quality security and governance in all with data scientists analysts and engineers to implement scalable and troubleshoot data pipelines in and guides team in estimating work necessary to realize a story/requirement through the software delivery solutions are well designed with maintainability/ease of integration and testing builtin from the for developing and delivering complex software requirements to accomplish business that software is developed to meet functional nonfunctional and compliance solutions Unit testing and ensure the solution can be integrated successfully into the overall application/system with clear robust and welltested Skills:5 years of handson experience in designing building and supporting Data Applications using Spark Sqoop and HiveBachelor or masters degree in computer science or related fieldStrong knowledge of working with large data sets and highcapacity big data processing platformStrong experience in Unix and Shell scriptingAdvanced knowledge of the Hadoop ecosystem and its componentsIndepth knowledge of Hive Shell scripting Python SparkAbility to write MapReduce using Job Schedulers like AutosysHands on experience in HiveQLGood knowledge on Hadoop Architecture and HDFSStrong knowledge of working with large data sets and highcapacity big data processing platformStrong experience in Unix and Shell scriptingExperience with Jenkins for Continuous IntegrationExperience using Source Code and Version Control Systems like Bitbucket GitGood to have experience on Agile DevelopmentRequired Skills: Hadoop Hive HDFS Spark Python UnixNice to Have Skills: Experience in international and multicultural environment AML Domain knowledge