Tittle : Snowflake Data Engineer
Location: Dallas TX (3 days a week is must)
Key Responsibilities
- Design and implement Snowflake schemas (star snowflake data vault) optimized with micro-partitioning clustering keys materialized views and search optimization services.
- Build real-time and batch ingestion pipelines into Snowflake using Snowpipe Kafka Connect Fivetran Matillion Informatica or dbt.
- Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
- Use Zero-Copy Cloning for environment management testing and sandboxing.
- Apply Time Travel and Fail-safe features for data recovery and auditing.
- Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
- Design integrations with cloud storage (S3 Azure ADLS GCS) for staging and external tables.
- Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
- Enable semi-structured data handling (JSON Avro Parquet ORC XML) using VARIANT columns and lateral flattening.
- Integrate Snowflake with BI tools (Power BI Tableau) via live connections and semantic layers.
- Implement RBAC (Role-Based Access Control) Row Access Policies and Dynamic Data Masking for data security.
- Optimize compute usage with multi-cluster warehouses resource monitors and query performance tuning.
- Manage cost optimization strategies (warehouse auto-suspend query profiling storage/compute separation).
- Integrate with data catalog & governance platforms (Collibra Alation Informatica CDGC) using Snowflake metadata and APIs.
- Work with domain teams to deliver data products leveraging Snowflakes data mesh-friendly features.
- Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
- Support CI/CD automation for Snowflake code deployment using GitHub Actions Azure DevOps or dbt Cloud.
Qualifications
Education: Bachelors or Masters in Computer Science Data Engineering or related field.
Experience:
- 10 years of data engineering experience with 5 years in Snowflake Data Cloud.
- Expertise in SQL optimization and Snowflake performance tuning.
- Hands-on with Snowpipe Streams & Tasks Snowpark Zero-Copy Cloning and Secure Data Sharing.
- Proficiency in Python Scala or Java for Snowpark development.
- Experience integrating with cloud platforms like AWS.
- Exposure to ETL/ELT tools (Informatica Matillion Fivetran).
- Familiarity with CI/CD Git DevOps practices for data operations.
- Preferred Certifications:
- SnowPro Core Certification
Key Skills
- Snowflake-native feature design and implementation (Snowpark Streams Time Travel Secure Data Sharing)
- Data ingestion (Snowpipe CDC Kafka Fivetran)
- Semi-structured data handling (VARIANT JSON Avro Parquet)
- Advanced SQL and performance tuning
- Data governance (RBAC masking lineage catalogs)
- Cloud data platform integrations (AWS S3 Azure ADLS GCP GCS)
- BI and analytics tool integration
- Cost optimization and warehouse orchestration
Tittle : Snowflake Data Engineer Location: Dallas TX (3 days a week is must) Key Responsibilities Design and implement Snowflake schemas (star snowflake data vault) optimized with micro-partitioning clustering keys materialized views and search optimization services. Build real-time and batch inges...
Tittle : Snowflake Data Engineer
Location: Dallas TX (3 days a week is must)
Key Responsibilities
- Design and implement Snowflake schemas (star snowflake data vault) optimized with micro-partitioning clustering keys materialized views and search optimization services.
- Build real-time and batch ingestion pipelines into Snowflake using Snowpipe Kafka Connect Fivetran Matillion Informatica or dbt.
- Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
- Use Zero-Copy Cloning for environment management testing and sandboxing.
- Apply Time Travel and Fail-safe features for data recovery and auditing.
- Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
- Design integrations with cloud storage (S3 Azure ADLS GCS) for staging and external tables.
- Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
- Enable semi-structured data handling (JSON Avro Parquet ORC XML) using VARIANT columns and lateral flattening.
- Integrate Snowflake with BI tools (Power BI Tableau) via live connections and semantic layers.
- Implement RBAC (Role-Based Access Control) Row Access Policies and Dynamic Data Masking for data security.
- Optimize compute usage with multi-cluster warehouses resource monitors and query performance tuning.
- Manage cost optimization strategies (warehouse auto-suspend query profiling storage/compute separation).
- Integrate with data catalog & governance platforms (Collibra Alation Informatica CDGC) using Snowflake metadata and APIs.
- Work with domain teams to deliver data products leveraging Snowflakes data mesh-friendly features.
- Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
- Support CI/CD automation for Snowflake code deployment using GitHub Actions Azure DevOps or dbt Cloud.
Qualifications
Education: Bachelors or Masters in Computer Science Data Engineering or related field.
Experience:
- 10 years of data engineering experience with 5 years in Snowflake Data Cloud.
- Expertise in SQL optimization and Snowflake performance tuning.
- Hands-on with Snowpipe Streams & Tasks Snowpark Zero-Copy Cloning and Secure Data Sharing.
- Proficiency in Python Scala or Java for Snowpark development.
- Experience integrating with cloud platforms like AWS.
- Exposure to ETL/ELT tools (Informatica Matillion Fivetran).
- Familiarity with CI/CD Git DevOps practices for data operations.
- Preferred Certifications:
- SnowPro Core Certification
Key Skills
- Snowflake-native feature design and implementation (Snowpark Streams Time Travel Secure Data Sharing)
- Data ingestion (Snowpipe CDC Kafka Fivetran)
- Semi-structured data handling (VARIANT JSON Avro Parquet)
- Advanced SQL and performance tuning
- Data governance (RBAC masking lineage catalogs)
- Cloud data platform integrations (AWS S3 Azure ADLS GCP GCS)
- BI and analytics tool integration
- Cost optimization and warehouse orchestration
View more
View less