Total IT / development experience of 3 years
Experience in Spark (Scala-Spark or PySpark) developing Big Data applications on Hadoop Hive and/or Kafka HBase MongoDBB4A3:B5d technology strategies
Exposure to deploying on Cloud platforms
At least 2 years of development experience on designing and developing Data Pipelines for Data Ingestion or Transformation using Spark-Scala/PySpark
At least 2 years of development experience in the following Big Data frameworks: File Format (Parquet AVRO ORC) Resource Management Distributed Processing and RDBMS
At least 2 years of developing applications in Agile with Monitoring Build Tools Version Control Unit Test Unix Shell Scripting TDD CI/CD Change Management to support DevOps
GOOD-TO-HAVE Banking domain knowledge
Hands-on experience in SAS toolset / statistical modelling migrating to Machine Learning models
Banking Risk Fraud or Digital Marketing Machine Learning models and use cases
ETL / Data Warehousing SQL and Data Modelling experience prior to Big Data experience Location Chennai / Bangalore / Pune Experience 4-7 Yrs
pyspark,ci/cd,machine learning,mongodb,data pipelines,unit test,data,unix shell scripting,hadoop,hive,kafka,learning,agile,spark,management,scala-spark,change management,machine learning models,a3,hbase,tdd,scala,big data