Our client is a digital services provider operating within the iGaming field. As part of their growth and expansion they are now seeking to recruit an ML Engineer (LLM / Google Cloud) who will be responsible for training and Fine-tuning text models (LLMs) deploying them on Google Cloud and building automation around these models.
The core mission: take example texts train the model so that the output strictly follows the required format and build reliable infrastructure and services that will call this model in production.
Tasks
Analyse business requirements for the desired output format and the logic the model must implement.
Prepare datasets based on example texts: cleaning annotation creating training/validation splits.
Train and fine-tune LLMs for specific use cases:
configure training parameters;
experiment with prompts system instructions input/output formats.
Evaluate model quality:
design and track metrics;
create test scenarios and A/B experiments;
ensure output format consistency and stability.
Deploy models to Google Cloud (for example via Vertex AI Cloud Run Kubernetes etc.).
Develop services and APIs (REST/gRPC) that expose the model to other systems.
Build automations and integrations that call the model:
background jobs queues event-driven triggers;
integration with internal services and databases.
Implement MLOps pipelines:
automate training / retraining workflows;
version models and datasets;
monitor model performance and quality in production.
Document models pipelines APIs and architectural decisions.
Requirements
3 years of software development experience (preferably Python).
Hands-on experience with ML / NLP: understanding of models loss functions training and validation workflows.
Practical experience with at least one ML framework: TensorFlow PyTorch Hugging Face etc.
Experience with Google Cloud:
Core services (Cloud Storage IAM VPC);
ideally Vertex AI Cloud Run Pub/Sub or similar.
Experience deploying models into production (API services containerization with Docker CI/CD).
Experience building and integrating REST APIs; confident working with JSON/JSONL logging and monitoring.
Understanding of how to design reliable and scalable systems (error handling retries queues timeouts).
Direct experience with LLMs: prompt engineering few-shot learning RAG.
Experience with MLOps tools (MLflow Vertex AI Pipelines or equivalents).
Experience with messaging/queue systems (Pub/Sub Kafka RabbitMQ) and workflow orchestration (Workflows Airflow etc.).
Understanding of data security and handling sensitive information including access control (IAM).
Benefits
- Highly competitive Salary along with quarterly bonuses
- Opportunity to work on fully remote basis under a B2B Service Agreement
Our client is a digital services provider operating within the iGaming field. As part of their growth and expansion they are now seeking to recruit an ML Engineer (LLM / Google Cloud) who will be responsible for training and Fine-tuning text models (LLMs) deploying them on Google Cloud and building ...
Our client is a digital services provider operating within the iGaming field. As part of their growth and expansion they are now seeking to recruit an ML Engineer (LLM / Google Cloud) who will be responsible for training and Fine-tuning text models (LLMs) deploying them on Google Cloud and building automation around these models.
The core mission: take example texts train the model so that the output strictly follows the required format and build reliable infrastructure and services that will call this model in production.
Tasks
Analyse business requirements for the desired output format and the logic the model must implement.
Prepare datasets based on example texts: cleaning annotation creating training/validation splits.
Train and fine-tune LLMs for specific use cases:
configure training parameters;
experiment with prompts system instructions input/output formats.
Evaluate model quality:
design and track metrics;
create test scenarios and A/B experiments;
ensure output format consistency and stability.
Deploy models to Google Cloud (for example via Vertex AI Cloud Run Kubernetes etc.).
Develop services and APIs (REST/gRPC) that expose the model to other systems.
Build automations and integrations that call the model:
background jobs queues event-driven triggers;
integration with internal services and databases.
Implement MLOps pipelines:
automate training / retraining workflows;
version models and datasets;
monitor model performance and quality in production.
Document models pipelines APIs and architectural decisions.
Requirements
3 years of software development experience (preferably Python).
Hands-on experience with ML / NLP: understanding of models loss functions training and validation workflows.
Practical experience with at least one ML framework: TensorFlow PyTorch Hugging Face etc.
Experience with Google Cloud:
Core services (Cloud Storage IAM VPC);
ideally Vertex AI Cloud Run Pub/Sub or similar.
Experience deploying models into production (API services containerization with Docker CI/CD).
Experience building and integrating REST APIs; confident working with JSON/JSONL logging and monitoring.
Understanding of how to design reliable and scalable systems (error handling retries queues timeouts).
Direct experience with LLMs: prompt engineering few-shot learning RAG.
Experience with MLOps tools (MLflow Vertex AI Pipelines or equivalents).
Experience with messaging/queue systems (Pub/Sub Kafka RabbitMQ) and workflow orchestration (Workflows Airflow etc.).
Understanding of data security and handling sensitive information including access control (IAM).
Benefits
- Highly competitive Salary along with quarterly bonuses
- Opportunity to work on fully remote basis under a B2B Service Agreement
View more
View less