Snowflake Data Engineer

Not Interested
Bookmark
Report This Job

profile Job Location:

Dallas, IA - USA

profile Monthly Salary: Not Disclosed
Posted on: 9 hours ago
Vacancies: 1 Vacancy

Job Summary

Tittle : Snowflake Data Engineer

Location: Dallas TX (3 days a week is must)

Key Responsibilities

  • Design and implement Snowflake schemas (star snowflake data vault) optimized with micro-partitioning clustering keys materialized views and search optimization services.
  • Build real-time and batch ingestion pipelines into Snowflake using Snowpipe Kafka Connect Fivetran Matillion Informatica or dbt.
  • Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
  • Use Zero-Copy Cloning for environment management testing and sandboxing.
  • Apply Time Travel and Fail-safe features for data recovery and auditing.
  • Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
  • Design integrations with cloud storage (S3 Azure ADLS GCS) for staging and external tables.
  • Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
  • Enable semi-structured data handling (JSON Avro Parquet ORC XML) using VARIANT columns and lateral flattening.
  • Integrate Snowflake with BI tools (Power BI Tableau) via live connections and semantic layers.
  • Implement RBAC (Role-Based Access Control) Row Access Policies and Dynamic Data Masking for data security.
  • Optimize compute usage with multi-cluster warehouses resource monitors and query performance tuning.
  • Manage cost optimization strategies (warehouse auto-suspend query profiling storage/compute separation).
  • Integrate with data catalog & governance platforms (Collibra Alation Informatica CDGC) using Snowflake metadata and APIs.
  • Work with domain teams to deliver data products leveraging Snowflakes data mesh-friendly features.
  • Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
  • Support CI/CD automation for Snowflake code deployment using GitHub Actions Azure DevOps or dbt Cloud.

Qualifications

Education: Bachelors or Masters in Computer Science Data Engineering or related field.

Experience:

  • 10 years of data engineering experience with 5 years in Snowflake Data Cloud.
  • Expertise in SQL optimization and Snowflake performance tuning.
  • Hands-on with Snowpipe Streams & Tasks Snowpark Zero-Copy Cloning and Secure Data Sharing.
  • Proficiency in Python Scala or Java for Snowpark development.
  • Experience integrating with cloud platforms like AWS.
  • Exposure to ETL/ELT tools (Informatica Matillion Fivetran).
  • Familiarity with CI/CD Git DevOps practices for data operations.
  • Preferred Certifications:
  • SnowPro Core Certification

Key Skills

  • Snowflake-native feature design and implementation (Snowpark Streams Time Travel Secure Data Sharing)
  • Data ingestion (Snowpipe CDC Kafka Fivetran)
  • Semi-structured data handling (VARIANT JSON Avro Parquet)
  • Advanced SQL and performance tuning
  • Data governance (RBAC masking lineage catalogs)
  • Cloud data platform integrations (AWS S3 Azure ADLS GCP GCS)
  • BI and analytics tool integration
  • Cost optimization and warehouse orchestration

Tittle : Snowflake Data Engineer Location: Dallas TX (3 days a week is must) Key Responsibilities Design and implement Snowflake schemas (star snowflake data vault) optimized with micro-partitioning clustering keys materialized views and search optimization services. Build real-time and batch inges...
View more view more

Key Skills

  • Apache Hive
  • S3
  • Hadoop
  • Redshift
  • Spark
  • AWS
  • Apache Pig
  • NoSQL
  • Big Data
  • Data Warehouse
  • Kafka
  • Scala