Design and deliver GenAI solutions including LLM/RAG (retrieval strategy, embeddings, vector stores, prompt flows, grounding) for enterprise use cases.
Evaluate and improve solution quality using offline/online metrics (quality, latency, cost) and iterate based on feedback.
Harden solutions for production with observability/monitoring, tracing, guardrails, safety controls, and reliability practices
Build and integrate model endpoints into products and workflows (APIs/services), partnering with engineering through to deployment.
Work across cloud platforms (Azure/AWS/GCP) integrating storage, compute, orchestration, and model/runtime components.
Assess data readiness for modelling/RAG (fitness, quality, access) and define remediation requirements
Collaborate in cross-functional squads (DS/DE/engineering/product) and contribute to reusable assets and ways of working.
Communicate clearly with stakeholders on trade-offs, evaluation results, risks, and adoption actions.
Own end-to-end workstream delivery, lead stakeholder conversations, mentor others. (more senior levels)
Shape solution direction and quality bar, coach teams, contribute to sales pursuits/bids and accelerators (most senior levels)
Requirements
Essential Skills:
Strong Python/R (pandas/NumPy; ML libs such as scikit-learn; DL frameworks TensorFlow/PyTorch).
Experience with LLM/RAG toolchains (e.g., LangChain, LlamaIndex, Semantic Kernel) and vector search (e.g., Pinecone, Weaviate, FAISS, Azure AI Search).
Experience with GenAI platforms (e.g., OpenAI API, Anthropic, Gemini, Llama or equivalents).
Exposure to big data/distributed computing and pipeline/feature engineering.