About Koantek:
Koantek is a Databricks Pure-Play Elite Partner helping enterprises modernize faster and unlock the full power of Data and AI. Backed by Databricks Ventures and honored as a six-time Databricks Partner of the Year we enable global enterprises to modernize at speed operationalize AI and realize the full value of their data. Our deep expertise spans industries such as healthcare financial services retail and SaaS delivering end-to-end solutions from rapid prototyping to production-scale AI deployments. We deliver tailored solutions that enable businesses to leverage data for growth and innovation. Our team of experts utilizes deep industry knowledge combined with cutting-edge technologies tools and methodologies to drive impactful results. By partnering with clients across a diverse range of industriesfrom emerging startups to established enterpriseswe help them uncover new opportunities and achieve a competitive advantage in the digital age.
About the Role
As a Solutions Architect at Koantek you will collaborate with customers to design scalable data architectures utilizing Databricks technology and services. The RSA at Koantek builds secure highly scalable big data solutions to achieve tangible data-driven outcomes all the while keeping simplicity and operational effectiveness in mind. Leveraging your technical expertise and business acumen you will navigate complex technology discussions showcasing the value of the Databricks platform throughout the sales process. Working alongside Account Executives you will engage with customers technical leaders including architects engineers and operations teams aiming to become a trusted advisor who
delivers concrete outcomes. This role collaborates with teammates product teams and cross-functional project teams to lead the adoption and integration of the Databricks Platform into the enterprise ecosystem and AWS/Azure/GCP architecture.
The impact you will have:
Develop Account Strategies: Work with Sales and other essential partners to
develop strategies for your assigned accounts to grow their usage of the Databricks
platform.
Establish Architecture Standards: Establish the Databricks Lakehouse architecture as the standard data architecture for customers through excellent technical account planning.
Demonstrate Value: Build and present reference architectures and demo applications to help prospects understand how Databricks can be used to achieve their goals and land new use cases.
Capture Technical Wins: Consult on big data architectures data engineering pipelines and data science/machine learning projects to prove out Databricks technology for strategic customer projects. Validate integrations with cloud services
and other third-party applications.
Promote Open-Source Projects: Become an expert in and promote Databricks-
inspired open-source projects (Spark Delta Lake MLflow) across developer
communities through meetups conferences and webinars.
Requirements
Technical Expertise:
Experience translating a customers business needs to technology solutions including establishing buy-in with essential customer stakeholders at all levels of the business.
Experienced at designing architecting and presenting data systems for customers and managing the delivery of production solutions of those data architectures.
Projects delivered with hands-on experience in development on databricks
Expert-level knowledge of data frameworks data lakes and open-source projects such as Apache Spark MLflow and Delta Lake
Expert-level hands-on coding experience in Spark/Scala Python or Pyspark
In depth understanding of Spark Architecture including Spark Core Spark SQL Data Frames Spark Streaming RDD caching Spark MLibT/event-driven/microservices in the cloud
Deep experience with distributed computing with spark with knowledge of spark runtime
Experience with private and public cloud architectures pros/cons and migration considerations.
Extensive hands-on experience implementing data migration and data processing using AWS/Azure/GCP services
Familiarity with CI/CD for production deployments
Familiarity with optimization for performance and scalability
Completed data engineering professional certification and required classes
SQL Proficiency: Fluent in SQL and database technology
Educational Background
Degree in a quantitative discipline (Computer Science Applied Mathematics Operations Research.
Relevant certifications (e.g. Databricks certifications AWS/Azure/GCP AI/ML certifications) are a plus. Workplace Flexibility
On-site presence at customer locations MAY be required based on the project and business needs. Candidates should be willing and able to travel for short or medium-term assignments when necessary.
About Koantek:Koantek is a Databricks Pure-Play Elite Partner helping enterprises modernize faster and unlock the full power of Data and AI. Backed by Databricks Ventures and honored as a six-time Databricks Partner of the Year we enable global enterprises to modernize at speed operationalize AI and...
About Koantek:
Koantek is a Databricks Pure-Play Elite Partner helping enterprises modernize faster and unlock the full power of Data and AI. Backed by Databricks Ventures and honored as a six-time Databricks Partner of the Year we enable global enterprises to modernize at speed operationalize AI and realize the full value of their data. Our deep expertise spans industries such as healthcare financial services retail and SaaS delivering end-to-end solutions from rapid prototyping to production-scale AI deployments. We deliver tailored solutions that enable businesses to leverage data for growth and innovation. Our team of experts utilizes deep industry knowledge combined with cutting-edge technologies tools and methodologies to drive impactful results. By partnering with clients across a diverse range of industriesfrom emerging startups to established enterpriseswe help them uncover new opportunities and achieve a competitive advantage in the digital age.
About the Role
As a Solutions Architect at Koantek you will collaborate with customers to design scalable data architectures utilizing Databricks technology and services. The RSA at Koantek builds secure highly scalable big data solutions to achieve tangible data-driven outcomes all the while keeping simplicity and operational effectiveness in mind. Leveraging your technical expertise and business acumen you will navigate complex technology discussions showcasing the value of the Databricks platform throughout the sales process. Working alongside Account Executives you will engage with customers technical leaders including architects engineers and operations teams aiming to become a trusted advisor who
delivers concrete outcomes. This role collaborates with teammates product teams and cross-functional project teams to lead the adoption and integration of the Databricks Platform into the enterprise ecosystem and AWS/Azure/GCP architecture.
The impact you will have:
Develop Account Strategies: Work with Sales and other essential partners to
develop strategies for your assigned accounts to grow their usage of the Databricks
platform.
Establish Architecture Standards: Establish the Databricks Lakehouse architecture as the standard data architecture for customers through excellent technical account planning.
Demonstrate Value: Build and present reference architectures and demo applications to help prospects understand how Databricks can be used to achieve their goals and land new use cases.
Capture Technical Wins: Consult on big data architectures data engineering pipelines and data science/machine learning projects to prove out Databricks technology for strategic customer projects. Validate integrations with cloud services
and other third-party applications.
Promote Open-Source Projects: Become an expert in and promote Databricks-
inspired open-source projects (Spark Delta Lake MLflow) across developer
communities through meetups conferences and webinars.
Requirements
Technical Expertise:
Experience translating a customers business needs to technology solutions including establishing buy-in with essential customer stakeholders at all levels of the business.
Experienced at designing architecting and presenting data systems for customers and managing the delivery of production solutions of those data architectures.
Projects delivered with hands-on experience in development on databricks
Expert-level knowledge of data frameworks data lakes and open-source projects such as Apache Spark MLflow and Delta Lake
Expert-level hands-on coding experience in Spark/Scala Python or Pyspark
In depth understanding of Spark Architecture including Spark Core Spark SQL Data Frames Spark Streaming RDD caching Spark MLibT/event-driven/microservices in the cloud
Deep experience with distributed computing with spark with knowledge of spark runtime
Experience with private and public cloud architectures pros/cons and migration considerations.
Extensive hands-on experience implementing data migration and data processing using AWS/Azure/GCP services
Familiarity with CI/CD for production deployments
Familiarity with optimization for performance and scalability
Completed data engineering professional certification and required classes
SQL Proficiency: Fluent in SQL and database technology
Educational Background
Degree in a quantitative discipline (Computer Science Applied Mathematics Operations Research.
Relevant certifications (e.g. Databricks certifications AWS/Azure/GCP AI/ML certifications) are a plus. Workplace Flexibility
On-site presence at customer locations MAY be required based on the project and business needs. Candidates should be willing and able to travel for short or medium-term assignments when necessary.