Detailed Job Description:
We are looking for a dedicated Hadoop Big Data Engineer to manage one of the largest distributed database infrastructures. The successful candidate will enjoy Development, Building, and Managing of highly available and efficient distributed database systems that serve around the globe.
Key Qualifications:
5+ Years of strong experience in developing Big Data applications using Scala/ Java/ Python, Spark, Hadoop, HDFS, Hive, Oozie, Kafka, and Map Reduce is a huge plus
Programming experience in building high-quality applications, data pipelines, and analytics solutions.
Design and build highly scalable data pipelines using new generation tools and technologies like Spark, Kafka to induct data from various distributed database systems.
Experience building large-scale data pipelines using AWS S3, EMR, Dynamodb, Lambda, Spark
Experience in designing and building dimensional data models to improve accessibility, efficiency, and quality of data
Strong analytical and communication skills. Should be self-driven, highly motivated, and able to learn quick
Must have:
Experience in working on Hadoop Distribution (CDH/HDP/MapR).
Hands-on experience with MapReduce, Hive 2.x*, Spark 2.x*.
Conceptual knowledge of Data Structures & Algorithms
Possessing in-depth knowledge of various Design Patterns, Data Processing Patterns
Knowledge and experience with NoSQL Database
(Cassandra/HBase/MongoDB/CouchDB/Neo4j),SQL Database(MySQL/Oracle).
Programming experience with Python/Java
Excellent communication, problem-solving & analytical skills with ability to thrive in a fast paced, dynamic environment & operate under stringent deadlines