REQUIREMENTS:
- Total experience 7.5 years
- Strong hands-on expertise in LLM engineering and Python backend development.
- Expertise in LLM Application Frameworks Prompt Engineering with LLMs Python FastAP
- Proven experience building and deploying applications using cuttingedge LLMs (GPT4/5 Claude Gemini Mistral LLaMA Mixtral DeepSeek etc.).
- Strong experience with RAG pipelines embeddings prompt engineering and multiagent systems.
- Hands-on expertise with LLM frameworks such as LangChain LlamaIndex Haystack DSPy AutoGen CrewAI.
- Deep knowledge of model finetuning techniques such as LoRA QLoRA PEFT adapters.
- Experience deploying opensource LLMs using vLLM TGI Ollama LM Studio Triton etc.
- Strong backend engineering experience with FastAPI (expert) Django or Flask microservices and distributed systems.
- Experience implementing REST GraphQL and streaming APIs.
- Hands-on experience with vector databases such as Pinecone Weaviate Milvus Qdrant FAISS Chroma.
- Knowledge of semantic search hybrid search embedding pipelines and enterprise knowledge systems.
- Strong understanding of cloud platforms (AWS GCP Azure) containers and Kubernetes.
- Experience with MLOps/LLMOps practicesCI/CD for ML workflows monitoring logging tracing and model lifecycle management.
- Bachelors/Masters in CS AI Data Science or equivalent experience.
- Excellent communication collaboration and problem-solving skills.
RESPONSIBILITIES:
- Design implement and optimize LLM-powered applications using leading and opensource models.
- Develop advanced prompt engineering system prompts and structured output pipelines.
- Build RAG pipelines with hybrid search embeddings and custom retrieval strategies.
- Develop multi-agent systems and autonomous AI workflows.
- Finetune adapt and serve foundation models using LoRA/QLoRA and modern inference engines.
- Deploy and scale LLM workloads using vLLM TGI Ollama or GPU/TPU-based systems.
- Integrate multimodal models across text image audio and video.
- Build evaluation pipelines for hallucination detection factual accuracy quality scoring and alignment.
- Implement guardrails moderation and safety policies for AI systems.
- Build scalable backend systems using FastAPI microservices event-driven architectures and secure API frameworks.
- Optimize backend performance observability and reliability.
- Build ingestion pipelines for document processing chunking preprocessing and semantic indexing.
- Implement semantic vector and hybrid search at scale.
- Deploy AI systems on cloud platforms manage Kubernetes inference clusters and optimize GPU utilization.
- Set up CI/CD automated testing model versioning and production monitoring for AI workflows.
- Develop enterprise-grade search knowledge systems and document intelligence platforms.
- Ensure robustness security and scalability in all AI and backend systems.
- Stay updated with the latest GenAI LLMOps and backend engineering innovations and share knowledge within the technical community.
Qualifications :
Bachelors or masters degree in computer science Information Technology or a related field.
Remote Work :
Yes
Employment Type :
Full-time
REQUIREMENTS:Total experience 7.5 yearsStrong hands-on expertise in LLM engineering and Python backend development.Expertise in LLM Application Frameworks Prompt Engineering with LLMs Python FastAPProven experience building and deploying applications using cuttingedge LLMs (GPT4/5 Claude Gemini Mist...
REQUIREMENTS:
- Total experience 7.5 years
- Strong hands-on expertise in LLM engineering and Python backend development.
- Expertise in LLM Application Frameworks Prompt Engineering with LLMs Python FastAP
- Proven experience building and deploying applications using cuttingedge LLMs (GPT4/5 Claude Gemini Mistral LLaMA Mixtral DeepSeek etc.).
- Strong experience with RAG pipelines embeddings prompt engineering and multiagent systems.
- Hands-on expertise with LLM frameworks such as LangChain LlamaIndex Haystack DSPy AutoGen CrewAI.
- Deep knowledge of model finetuning techniques such as LoRA QLoRA PEFT adapters.
- Experience deploying opensource LLMs using vLLM TGI Ollama LM Studio Triton etc.
- Strong backend engineering experience with FastAPI (expert) Django or Flask microservices and distributed systems.
- Experience implementing REST GraphQL and streaming APIs.
- Hands-on experience with vector databases such as Pinecone Weaviate Milvus Qdrant FAISS Chroma.
- Knowledge of semantic search hybrid search embedding pipelines and enterprise knowledge systems.
- Strong understanding of cloud platforms (AWS GCP Azure) containers and Kubernetes.
- Experience with MLOps/LLMOps practicesCI/CD for ML workflows monitoring logging tracing and model lifecycle management.
- Bachelors/Masters in CS AI Data Science or equivalent experience.
- Excellent communication collaboration and problem-solving skills.
RESPONSIBILITIES:
- Design implement and optimize LLM-powered applications using leading and opensource models.
- Develop advanced prompt engineering system prompts and structured output pipelines.
- Build RAG pipelines with hybrid search embeddings and custom retrieval strategies.
- Develop multi-agent systems and autonomous AI workflows.
- Finetune adapt and serve foundation models using LoRA/QLoRA and modern inference engines.
- Deploy and scale LLM workloads using vLLM TGI Ollama or GPU/TPU-based systems.
- Integrate multimodal models across text image audio and video.
- Build evaluation pipelines for hallucination detection factual accuracy quality scoring and alignment.
- Implement guardrails moderation and safety policies for AI systems.
- Build scalable backend systems using FastAPI microservices event-driven architectures and secure API frameworks.
- Optimize backend performance observability and reliability.
- Build ingestion pipelines for document processing chunking preprocessing and semantic indexing.
- Implement semantic vector and hybrid search at scale.
- Deploy AI systems on cloud platforms manage Kubernetes inference clusters and optimize GPU utilization.
- Set up CI/CD automated testing model versioning and production monitoring for AI workflows.
- Develop enterprise-grade search knowledge systems and document intelligence platforms.
- Ensure robustness security and scalability in all AI and backend systems.
- Stay updated with the latest GenAI LLMOps and backend engineering innovations and share knowledge within the technical community.
Qualifications :
Bachelors or masters degree in computer science Information Technology or a related field.
Remote Work :
Yes
Employment Type :
Full-time
View more
View less