Job Title: DevOps Engineer
Location: Boston MA
Work Type: Onsite Role
Duration: 12 Months (Temporary)
Work Schedule: Monday-Friday 9:00 AM - 5:00 PM ESTNOTE: NO H1B // Local Candidates with DLPrimary Skills:DevOps Engineering
Snowflake (Roles Warehouses Performance Tuning Cost Control RBAC)
Informatica Intelligent Cloud Services (IICS)
Apache Airflow (DAG Orchestration Monitoring Deployment)
CI/CD Pipelines (GitLab GitHub Actions Azure DevOps Jenkins)
Terraform (Infrastructure as Code - IaC)
Cloud Platforms: AWS Azure GCP
Docker & Kubernetes
Python & Shell Scripting
DevSecOps Security Monitoring & Automation
Position OverviewThe client is seeking an experienced DevOps Engineer to support a cloud data warehouse modernization initiative transitioning from a SQL Server/AWS-based system to a Snowflake-based data platform. This role is responsible for developing maintaining and optimizing data pipelines and integration processes that support analytics reporting and business operations.
The DevOps Engineer will design and implement CI/CD pipelines automate data pipeline deployments and ensure high availability security and operational reliability across Snowflake Informatica (IICS) and Apache Airflow environments.
Job Duties & ResponsibilitiesBuild and maintain CI/CD pipelines for Snowflake Informatica (IICS) and Apache Airflow DAG deployments
Implement automated code promotion across development test and production environments
Integrate testing linting and security scanning into deployment workflows
Develop and manage Infrastructure as Code (IaC) using Terraform or similar tools
Ensure configuration and environment consistency across multi-region and multi-cloud setups
Maintain secure connectivity between cloud and on-prem systems (VPNs private links firewalls)
Implement logging and alerting for Airflow DAGs Informatica workflows and Snowflake performance
Build proactive monitoring dashboards for job failures data quality checks and warehouse usage
Optimize Snowflake pipeline performance concurrency and cost governance
Own deployment frameworks for ETL/ELT code SQL scripts and metadata updates
Support user access provisioning and RBAC alignment across Snowflake Informatica and Airflow
Troubleshoot platform and orchestration issues and lead incident response during outages
Enforce DevSecOps best practices including encryption secrets management and key rotation
Implement audit logging compliance and backup/restore strategies aligned with governance requirements
Participate in testing deployment and release management for new data workflows and enhancements
Required Skills Experience Qualifications & AbilitiesBachelors degree or equivalent experience in Computer Science Information Systems Data Engineering Health Informatics or a related field
3-7 years of experience in DevOps Cloud Engineering or Data Platform Engineering
Strong hands-on experience with Snowflake (roles warehouses performance optimization cost governance)
Expertise in Apache Airflow including DAG orchestration monitoring and deployments
Experience with Informatica (IICS) pipeline deployment automation
Strong CI/CD experience using GitLab GitHub Actions Azure DevOps Jenkins or similar tools
Proficiency in Terraform Python and Shell scripting
Deep understanding of cloud platforms such as AWS Azure or GCP
Experience with containerization technologies (Docker Kubernetes) particularly for Airflow workloads
Solid knowledge of networking concepts and security controls
Ability to troubleshoot complex distributed data workloads
Strong communication skills for collaboration with both technical and non-technical stakeholders
Excellent documentation and cross-team collaboration skills
Proactive mindset with a focus on automation and continuous process improvement
Detail-oriented approach with emphasis on data accuracy and operational excellence
Preferred Skills Experience & QualificationsExperience migrating from SQL Server or other legacy data warehouse platforms
Knowledge of FinOps practices for Snowflake cost and usage optimization
Background in healthcare finance or other regulated industries is a plus
Arushi Khanna Associate - Hiring & Recruitment
Email:
&
Vishal (Victor) Verma Assistant Manager
NS IT Solutions
Required Skills:
KUBERNETESCLOUD SERVICESAZUREGITHUBDEVOPSCOST CONTROLSQL SERVERAPACHE AIRFLOWGITLABDOCKERPYTHONCI/CDTERRAFORMAWSINFRASTRUCTURE AS CODEINFORMATICAAUTOMATIONTESTING
Job Title: DevOps EngineerLocation: Boston MAWork Type: Onsite RoleDuration: 12 Months (Temporary)Work Schedule: Monday-Friday 9:00 AM - 5:00 PM ESTNOTE: NO H1B // Local Candidates with DLPrimary Skills:DevOps EngineeringSnowflake (Roles Warehouses Performance Tuning Cost Control RBAC)Informatica In...
Job Title: DevOps Engineer
Location: Boston MA
Work Type: Onsite Role
Duration: 12 Months (Temporary)
Work Schedule: Monday-Friday 9:00 AM - 5:00 PM ESTNOTE: NO H1B // Local Candidates with DLPrimary Skills:DevOps Engineering
Snowflake (Roles Warehouses Performance Tuning Cost Control RBAC)
Informatica Intelligent Cloud Services (IICS)
Apache Airflow (DAG Orchestration Monitoring Deployment)
CI/CD Pipelines (GitLab GitHub Actions Azure DevOps Jenkins)
Terraform (Infrastructure as Code - IaC)
Cloud Platforms: AWS Azure GCP
Docker & Kubernetes
Python & Shell Scripting
DevSecOps Security Monitoring & Automation
Position OverviewThe client is seeking an experienced DevOps Engineer to support a cloud data warehouse modernization initiative transitioning from a SQL Server/AWS-based system to a Snowflake-based data platform. This role is responsible for developing maintaining and optimizing data pipelines and integration processes that support analytics reporting and business operations.
The DevOps Engineer will design and implement CI/CD pipelines automate data pipeline deployments and ensure high availability security and operational reliability across Snowflake Informatica (IICS) and Apache Airflow environments.
Job Duties & ResponsibilitiesBuild and maintain CI/CD pipelines for Snowflake Informatica (IICS) and Apache Airflow DAG deployments
Implement automated code promotion across development test and production environments
Integrate testing linting and security scanning into deployment workflows
Develop and manage Infrastructure as Code (IaC) using Terraform or similar tools
Ensure configuration and environment consistency across multi-region and multi-cloud setups
Maintain secure connectivity between cloud and on-prem systems (VPNs private links firewalls)
Implement logging and alerting for Airflow DAGs Informatica workflows and Snowflake performance
Build proactive monitoring dashboards for job failures data quality checks and warehouse usage
Optimize Snowflake pipeline performance concurrency and cost governance
Own deployment frameworks for ETL/ELT code SQL scripts and metadata updates
Support user access provisioning and RBAC alignment across Snowflake Informatica and Airflow
Troubleshoot platform and orchestration issues and lead incident response during outages
Enforce DevSecOps best practices including encryption secrets management and key rotation
Implement audit logging compliance and backup/restore strategies aligned with governance requirements
Participate in testing deployment and release management for new data workflows and enhancements
Required Skills Experience Qualifications & AbilitiesBachelors degree or equivalent experience in Computer Science Information Systems Data Engineering Health Informatics or a related field
3-7 years of experience in DevOps Cloud Engineering or Data Platform Engineering
Strong hands-on experience with Snowflake (roles warehouses performance optimization cost governance)
Expertise in Apache Airflow including DAG orchestration monitoring and deployments
Experience with Informatica (IICS) pipeline deployment automation
Strong CI/CD experience using GitLab GitHub Actions Azure DevOps Jenkins or similar tools
Proficiency in Terraform Python and Shell scripting
Deep understanding of cloud platforms such as AWS Azure or GCP
Experience with containerization technologies (Docker Kubernetes) particularly for Airflow workloads
Solid knowledge of networking concepts and security controls
Ability to troubleshoot complex distributed data workloads
Strong communication skills for collaboration with both technical and non-technical stakeholders
Excellent documentation and cross-team collaboration skills
Proactive mindset with a focus on automation and continuous process improvement
Detail-oriented approach with emphasis on data accuracy and operational excellence
Preferred Skills Experience & QualificationsExperience migrating from SQL Server or other legacy data warehouse platforms
Knowledge of FinOps practices for Snowflake cost and usage optimization
Background in healthcare finance or other regulated industries is a plus
Arushi Khanna Associate - Hiring & Recruitment
Email:
&
Vishal (Victor) Verma Assistant Manager
NS IT Solutions
Required Skills:
KUBERNETESCLOUD SERVICESAZUREGITHUBDEVOPSCOST CONTROLSQL SERVERAPACHE AIRFLOWGITLABDOCKERPYTHONCI/CDTERRAFORMAWSINFRASTRUCTURE AS CODEINFORMATICAAUTOMATIONTESTING
View more
View less