Job DescriptionToday the corporate landscape is dynamic and the world ahead is full of possibilities! None of the amazing things we do at Infosys would be possible without an equally amazing culture the environment where ideas can flourish and where you are empowered to move forward as far as your ideas will take you.At Infosys we assure that your career will never stand still we will inspire you to build whats next and we will navigate further together. Our journey of learnability values and trusted relationships with our clients continue to be the cornerstones of our organization and these values are upheld only because of our people.Your role- Design develop and maintain scalable data pipelines using AWS services (e.g. Glue EMR Redshift S3 Lambda) or databricks.
- Build ETL processes to ingest transform and load data from various sources into data lake (S3) and data warehouse (Redshift).
- Optimize data storage and retrieval for performance and cost using AWS and databricks best practices.
- Collaborate with data scientists analysts and business stakeholders to understand data requirements.
- Implement data quality validation and governance solutions.
- Monitor troubleshoot and improve pipelines data workflows and infrastructure.
- Automate data pipeline deployments using CI/CD and Infrastructure as Code (e.g. CloudFormation Terraform) in AWS and Databricks assent bundles.
- Document data architecture processes and technical solutions.
- Consistently learn about new and updated services offered by AWS in data engineering field.
- Ensure security and compliance of data solutions within AWS and databricks.
- 8 years of experience in AWS and Databricks data engineering
- AWS S3 - Experience in using Buckets Versioning S3 event triggers
- AWS Glue Experience in building Jobs External Libraries Glue Catalog Triggers Crawlers
- AWS Redshift/Snowflake Good understanding of Sort and Dist keys external tables table design Hash keys
- AWS Cloudwatch EventBridge Athena Familiar with Cloudwatch logs EventBridge rules Data querying via Athena
- AWS Lambda- Experience in building Lambda functions and setting up Lambda Triggers Layers
- DevOps : Familiar with CICD concepts Code repository Jenkins IACs (like CloudFormation)
- Programming Skills
- Python: Proficiency in using Python for data processing Understanding of different libraries
- Spark : Proficiency in using Spark SQL for data processing and using DataFrames for data manipulations
- SQL: Proficiency in writing complex SQL queries (Subqueries Nested Queries CTEs Windows function Aggregation)
- Data Lake DWH and ETL concepts : Understanding of Data lake DWH Concepts like Star schema & SCD Types ETL concepts
- Advanced expertise in data architecture modeling and warehousing (star/snowflake schema normalization denormalization).
- Extensive experience with ETL/ELT pipeline design orchestration and automation.
- Knowledge of regulatory compliance (GDPR)
- Performance tuning and optimization for large-scale data systems
- Experience with data lakehouse architectures
Preferred- Exposure to lake house data mesh data fabric or modern data architectures
- Experience with orchestration tools
- SAP ECC and Salesforce data understanding
- Data as a domain knowledge
- AWS Bedrock Generative AI skills
- Team management
- Exposure to lakehouse data mesh data fabric or modern data architectures
PersonalBesides the professional qualifications of the candidates we place great importance in addition to various forms personality profile. These include:- High analytical skills
- A high degree of initiative and flexibility
- High customer orientation
- High quality awareness
- Excellent verbal and written communication skills
About InfosysInfosys is a global leader in next-generation digital services and consulting. We enable clients in 46 countries to navigate their digital transformation. With over three decades of experience in managing the systems and workings of global enterprises we expertly steer our clients through the many next of their digital journey. We do it by enabling the enterprise with an AI-powered core that helps prioritize the execution of change. We also empower the business with agile digital at scale to deliver unprecedented levels of performance and customer delight. Our always-on learning agenda drives their continuous improvement through building and transferring digital skills expertise and ideas from our innovation ecosystem.Visit to see how Infosys can help your enterprise navigate your next.All aspects of employment at Infosys are based on merit competence and performance. We are committed to embracing diversity and creating an inclusive environment for all employees. Infosys is proud to be an equal opportunity employer Required Experience:
Contract
Job DescriptionToday the corporate landscape is dynamic and the world ahead is full of possibilities! None of the amazing things we do at Infosys would be possible without an equally amazing culture the environment where ideas can flourish and where you are empowered to move forward as far as your i...
Job DescriptionToday the corporate landscape is dynamic and the world ahead is full of possibilities! None of the amazing things we do at Infosys would be possible without an equally amazing culture the environment where ideas can flourish and where you are empowered to move forward as far as your ideas will take you.At Infosys we assure that your career will never stand still we will inspire you to build whats next and we will navigate further together. Our journey of learnability values and trusted relationships with our clients continue to be the cornerstones of our organization and these values are upheld only because of our people.Your role- Design develop and maintain scalable data pipelines using AWS services (e.g. Glue EMR Redshift S3 Lambda) or databricks.
- Build ETL processes to ingest transform and load data from various sources into data lake (S3) and data warehouse (Redshift).
- Optimize data storage and retrieval for performance and cost using AWS and databricks best practices.
- Collaborate with data scientists analysts and business stakeholders to understand data requirements.
- Implement data quality validation and governance solutions.
- Monitor troubleshoot and improve pipelines data workflows and infrastructure.
- Automate data pipeline deployments using CI/CD and Infrastructure as Code (e.g. CloudFormation Terraform) in AWS and Databricks assent bundles.
- Document data architecture processes and technical solutions.
- Consistently learn about new and updated services offered by AWS in data engineering field.
- Ensure security and compliance of data solutions within AWS and databricks.
- 8 years of experience in AWS and Databricks data engineering
- AWS S3 - Experience in using Buckets Versioning S3 event triggers
- AWS Glue Experience in building Jobs External Libraries Glue Catalog Triggers Crawlers
- AWS Redshift/Snowflake Good understanding of Sort and Dist keys external tables table design Hash keys
- AWS Cloudwatch EventBridge Athena Familiar with Cloudwatch logs EventBridge rules Data querying via Athena
- AWS Lambda- Experience in building Lambda functions and setting up Lambda Triggers Layers
- DevOps : Familiar with CICD concepts Code repository Jenkins IACs (like CloudFormation)
- Programming Skills
- Python: Proficiency in using Python for data processing Understanding of different libraries
- Spark : Proficiency in using Spark SQL for data processing and using DataFrames for data manipulations
- SQL: Proficiency in writing complex SQL queries (Subqueries Nested Queries CTEs Windows function Aggregation)
- Data Lake DWH and ETL concepts : Understanding of Data lake DWH Concepts like Star schema & SCD Types ETL concepts
- Advanced expertise in data architecture modeling and warehousing (star/snowflake schema normalization denormalization).
- Extensive experience with ETL/ELT pipeline design orchestration and automation.
- Knowledge of regulatory compliance (GDPR)
- Performance tuning and optimization for large-scale data systems
- Experience with data lakehouse architectures
Preferred- Exposure to lake house data mesh data fabric or modern data architectures
- Experience with orchestration tools
- SAP ECC and Salesforce data understanding
- Data as a domain knowledge
- AWS Bedrock Generative AI skills
- Team management
- Exposure to lakehouse data mesh data fabric or modern data architectures
PersonalBesides the professional qualifications of the candidates we place great importance in addition to various forms personality profile. These include:- High analytical skills
- A high degree of initiative and flexibility
- High customer orientation
- High quality awareness
- Excellent verbal and written communication skills
About InfosysInfosys is a global leader in next-generation digital services and consulting. We enable clients in 46 countries to navigate their digital transformation. With over three decades of experience in managing the systems and workings of global enterprises we expertly steer our clients through the many next of their digital journey. We do it by enabling the enterprise with an AI-powered core that helps prioritize the execution of change. We also empower the business with agile digital at scale to deliver unprecedented levels of performance and customer delight. Our always-on learning agenda drives their continuous improvement through building and transferring digital skills expertise and ideas from our innovation ecosystem.Visit to see how Infosys can help your enterprise navigate your next.All aspects of employment at Infosys are based on merit competence and performance. We are committed to embracing diversity and creating an inclusive environment for all employees. Infosys is proud to be an equal opportunity employer Required Experience:
Contract
View more
View less