· Lead the design and optimization of RAG (Retrieval-Augmented Generation) pipelines.
· Oversee model training, fine-tuning, and evaluation for LLMs and embeddings.
· Architect AI inference services integrating with LangChain, Qdrant, and Elasticsearch.
· Define standards for model observability, data lineage, and output traceability.
· Collaborate with backend and infra teams for scalable AI service deployment.
· Mentor AI and Data Engineers; review code, models, and architecture proposals.
· Continuously evaluate new AI frameworks (Groq, Anthropic, Hugging Face, Voyage AI).
Must-have
· 6+ years in AI/ML engineering, including 2+ years of team leadership.
· Expertise in Python, PyTorch, LangChain, and LLM-based architectures.
· Deep understanding of semantic retrieval, embedding models, and vector search.
· Experience with data pipeline orchestration (Airflow, Prefect).
· Familiar with Docker/Kubernetes, AWS, and Azure ML environments.
· Strong leadership and mentoring experience
Nice-to-have
· Hands-on with multi-modal AI (text + image).
· Experience implementing LLM observability tools (Langfuse, Weights & Biases).
· Base salary: Minimum $3,500 (net)
· Dynamic startup environment with plenty of opportunities for growth and learning.
· Lead the AI core of a global platform scaling to 15+M users.
· Build trusted, explainable AI systems used across industries.
· Work directly with CTO and partner with data experts and architects.
· Shape the future of AI infrastructure for verified knowledge.