Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
We are seeking an experienced Data Scientist with specialized expertise in Generative AI. The ideal candidate will join our team to work on diverse customer projects, collaborating with businesses to solve unique challenges through data. The ideal candidate will also contribute to our core product, a cloud-agnostic advanced chatbot solution built on Large Language Models (LLMs), Retrieval-Augmented Generation (RAG), embeddings, and various API components.
Your primary responsibilities will include using your expertise in statistics, machine learning, and programming to transform complex data into actionable insights and strategic recommendations that deliver tangible business value. You will be instrumental in both implementation projects and consulting missions. Additionally, you will be participating in the development of Back-End components for GenAI solutions, rigorously test and troubleshoot the solution through iterative releases based on defined functional requirements, and selecting appropriate technologies, libraries, and open-source tools.
Key Responsibilities
- Work closely with customers to understand their data needs and deliver tailored solutions through data analysis, data cleaning, consultancy, and implementation missions.
- Perform exploratory data analysis (EDA) and build machine learning models to solve customer-specific data challenges.
- Develop data pipelines and ETL processes to prepare and clean data for analysis and model training.
- Participate in the development of GenAI applications in Python, focusing on back-end components and APIs.
- Integrate Python components with LLM frameworks, RAG pipelines, embeddings, OCR libraries and databases
- Evaluate and contribute in architectural decisions and technology selection
- Package and deploy containerized components within Kubernetes environments
- Troubleshoot and resolve issues related to code, configuration, performance, and infrastructure
Profile Specify
- Proven experience in data science, demonstrating a solid understanding of the full project lifecycle, from problem framing and data collection to model deployment and monitoring
- Experience working with large and complex datasets from various sources
- Proficient in Python and API development, with the ability to write clean, scalable, and modular code using modern software design patterns
- Deep theoretical and practical knowledge of Generative AI
- Familiarity with Cloud Platform, containerized environments, CI/CD pipelines, and agile development workflows
- Capable of translating complex technical concepts into clear, understandable language for non-technical stakeholders
- Strong commitment to project deadlines with the ability to deliver tested and stable solutions.
- Effective in collaborative environments, contributing to code reviews and shared responsibilities.
Experience
- 5+ years of professional experience as a Data Scientist or similar role
- Experience in a client-facing role, including reporting and managing project outcomes for customers
- Proven track record of developing and deploying machine learning models in a production setting
- Practical experience with LLMs, embeddings, APIs, and GenAI libraries in production environments.
- Solid experience in Python and developing RESTful APIs.
- Familiarity developing in public cloud environments (Azure, Google Cloud Platform)
- Familiarity with containerization and deployment using Docker, Kubernetes, and Harbor.
Technical Skills
- Data & Machine Learning: NumPy, TensorFlow, PyTorch, Apache Airflow, and a strong understanding of MLOps principles
- Databases: Vector, SQL and NoSQL databases, specifically PostgreSQL, SQL Server.
- Machine learning operations (MLOps) principles and tools for model deployment and management.
- Languages/Programming: Python (including asynchronous programming), Structured programming, Bash, SQL, Rest APIs
- Version Control: Git platforms such as GitLab, GitHub, or Azure DevOps.
- Frameworks/APIs: LLM Frameworks (VLLM, Ollama...), FastAPI/Flask, vectorDBs (ChromaDB, FAISS…) , LangChain, LangGraph...
- Infrastructure: Docker, Kubernetes, Harbor, Cloud Storage Buckets
Soft Skills
- Collaborative team player, open to feedback, with a growth mindset
- Excellent communication skills, able to clearly document and explain complex systems and code
- Curious and open to experimenting with emerging AI tools and technologies
- Strong problem-solving skills with a meticulous and detail-oriented development approach
- Proactive and self-driven, with a strong focus on continuous improvement and quality
- Able to perform under pressure and manage tight effectively
- Committed to delivering high-quality, stable, and well-tested solutions
- Fluent in spoken and written in English, French is a plus
Key Skills
Ranked by relevanceReady to apply?
Join Proximus Luxembourg and take your career to the next level!
Application takes less than 5 minutes