About Infinitive: Infinitive is a data and AI consultancy that helps clients modernize monetize and operationalize their data to generate lasting value. They pride themselves on their deep industry and technology expertise ensuring that they drive and sustain the adoption of new capabilities. Infinitive is committed to aligning their team with their clients culture ensuring a successful partnership by bringing the right mix of talent and skills for high return on investment. Infinitive has earned recognition as one of the Best Small Firms to Work For by Consulting Magazine receiving this accolade seven times most recently in 2024. They have also been honored as a Top Workplace by the Washington Post Best Places to Work by the Washington Business Journal and Best Places to Work by Virginia Business.
Job Summary: We are seeking a skilled Data Engineer to join our team and play a key role in designing building and maintaining robust data pipelines and platforms. The ideal candidate will have strong experience with Python AWS (Glue S3 Lambda CloudWatch) Databricks Apache Spark SQL Snowflake and DynamoDB. This role involves working with largescale data processing systems optimizing ETL/ELT workflows and ensuring the reliability scalability and security of data solutions.
Key Responsibilities:
Design develop and maintain scalable ETL/ELT pipelines using AWS Glue Databricks and Apache Spark.
Work with structured and semistructured data in Snowflake DynamoDB and S3.
Optimize and troubleshoot Spark jobs for performance and cost efficiency.
Develop and maintain Lambda functions to support realtime and batch data processing.
Implement data quality validation and monitoring using CloudWatch and other observability tools.
Design and optimize complex SQL queries for data transformation aggregation and reporting.
Collaborate with crossfunctional teams including Data Scientists Analysts and DevOps teams to support datadriven decisionmaking.
Implement best practices for data security governance and compliance in the cloud environment.
Automate data workflows and CI/CD pipelines using infrastructureascode and Gitbased version control.
Required Qualifications:
3 years of experience in Data Engineering or a similar role.
Strong proficiency in Python for data processing and automation.
Handson experience with AWS services including Glue S3 Lambda CloudWatch DynamoDB.
Expertise in Databricks and Apache Spark for largescale data processing.
Proficient in SQL for querying and manipulating structured data.
Experience working with Snowflake including schema design and performance tuning.
Knowledge of data lake and data warehouse architectures.
Familiarity with data security IAM roles and cloudbased authentication mechanisms.
Strong problemsolving and debugging skills in a cloudbased data environment.
Preferred Qualifications:
Experience with streaming data architectures using Kafka Kinesis or Pub/Sub.
Knowledge of orchestration tools such as Airflow Step Functions or MWAA.
Experience with Terraform or CloudFormation for infrastructure automation.
Understanding of data governance frameworks (e.g. Dataplex Alation or Unity Catalog).
Experience in CI/CD and DevOps practices for data engineering pipelines.
Handson experience with monitoring and logging tools including Splunk and NewRelic.
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.