Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
You won't just train models; you will design, build, and own the entire lifecycle of AI systems that leverage Large Language Models (LLMs), RAG, and Multimodal AI to solve our clients' most complex problems.
The Challenge: Owning the LLM Production Pipeline
This role demands a unique combination of applied machine learning, data engineering, and scalable software delivery, operating entirely within a robust MLOps/LLMOps framework .
🛠️ What You Will Be Doing (Key Responsibilities):
- Design & Deliver GenAI Solutions: Leading the implementation of LLM-based applications, including custom chatbots, advanced summarization services, and creative co-pilots
- Model Optimization & Steering: Applying cutting-edge techniques like fine-tuning, LoRA/PEFT, and RLHF to optimize model performance and ensure factual accuracy
- RAG System Architecture: Architecting and building high-performance Retrieval-Augmented Generation (RAG) pipelines, managing the full lifecycle of embeddings and context-aware retrieval
- Production Code: Writing robust, deployment-ready software in Python and TypeScript, delivering clean, modular code
- MLOps Automation: Implementing full LLMOps/MLOps using Docker, Kubernetes, and Terraform to automate CI/CD, deployment, and versioning across major cloud platforms (AWS/GCP/Azure)
- Data Grounding: Ensuring model intelligence is fresh and reliable by connecting to Airflow, dbt, and Kafka pipelines
- Governance & Safety: Embedding responsible AI practices, including hallucination control, bias mitigation, and auditability, into every deployed system
- Deep knowledge of transformers, diffusion models, and advanced prompt engineering
- Proven expertise in embeddings, vector databases, and designing context-aware retrieval pipelines
- Proficiency in Python, REST APIs, containerization, CI/CD, and model versioning
- Proficiency in Python and TypeScript; core ML development using PyTorch, TensorFlow, and the Hugging Face ecosystem
- Expertise with LangChain, LlamaIndex, and APIs from OpenAI/Anthropic
- Use of MLflow and Weights & Biases for experiment tracking and versioning, leveraging platforms like Vertex AI or SageMaker
- Mastery of Docker, Kubernetes, and Terraform for scalable deployment across major clouds (AWS, GCP, Azure)
- Familiarity with Airflow, dbt, Postgres, and Kafka for data ingestion and transformation
- 3-7 years in AI/ML, with a critical 1-3 years hands-on with LLMs or production GenAI applications
- Proven delivery of production-grade GenAI tools in agile, rapid-iteration environments
- BSc/MSc in Computer Science, Data Science, or a related technical field
- Experience with model evaluation, data privacy, and bias mitigation
- Hands-on experience with ChromaDB, Pinecone, Weaviate, and FAISS for retrieval and embedding management
Benefits & Growth Opportunities:
- Competitive salary
- Comprehensive health insurance
- Ongoing professional development
- Opportunity to work on cutting-edge AI projects
- Flexible working arrangements
- Career advancement opportunities in a rapidly growing AI company
Please note: If you fail to meet the required criteria in the screening questions, your application will not be progressed
Key Skills
Ranked by relevanceReady to apply?
Join Deeplight AI and take your career to the next level!
Application takes less than 5 minutes

