Responsible for designing implementing and managing data and machine learning solutions on Google Cloud Platform
Key Responsibilities:
- Design end-to-end data solutions including data ingestion storage processing and analysis pipelines as well as machine learning model development deployment and monitoring pipelines.
- Design and implement scalable secure and cost-optimized cloud infrastructure using GCP services like BigQuery Dataflow Dataproc Cloud Storage and Kubernetes Engine.
- Design and implement data models ensuring data consistency accuracy and accessibility for various applications and users.
- Establish MLOps practices enabling the automation of machine learning model training deployment and monitoring.
- Ensure that all data solutions adhere to security and compliance standards implementing access controls encryption and other security measures.
- Monitor and optimize the performance of data and machine learning systems ensuring they meet business requirements and SLAs.
- Develop and implement strategies for managing and optimizing cloud costs ensuring efficient resource utilization.
- They provide technical guidance and mentorship to other team members fostering a culture of best practices and continuous improvement.
Key Skills:
- 10 years of experience designing and developing production grade data architectures using google cloud data services and solutions
- Proficiency in BigQuery Dataflow Dataproc Cloud Storage pub-sub Kubernetes Engine and other relevant GCP services.
- Strong Experience with data warehousing ETL processes data modeling and data pipeline development.
- Strong hands on experience in Python and SQL
- Strong experience of model development deployment and monitoring using Vertex AI
- Good experience of LLM agents and agentic AI Agent Space and hands on RAG experience
- Experience with cloud computing concepts including infrastructure as code (IaC) scalability security and cost optimization.