Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
In this role, you will focus on designing and maintaining scalable data pipelines and supporting the development of intelligent systems. You will be dedicated to a team working on AI agents and the infrastructure that powers them, contributing to the development of enterprise-grade cloud solutions using the latest AI technologies.
This is an opportunity to gain experience with real-world business cases, actively build the company's knowledge base in the field, and grow your expertise at the intersection of data engineering and artificial intelligence.
Challenges you'll tackle:
- Develop and maintain ETL/ELT pipelines using PySpark in Azure Databricks, with SQL for data transformations and Python/Pandas for data manipulation, where applicable
- Design and implement data models for structured and unstructured data
- Work on NLP, AI/ML, and agentic networks to build intelligent solutions
- Develop and optimise machine learning models and integrate them into data pipelines
- Collaborate with Data Scientists and Engineers to implement data-driven solutions
- Work with Git and version control to manage code and data pipelines effectively
- Research and experiment with new AI/ML techniques and apply them to real-world business problems
Skills for success:
- 2+ years of experience in Data Engineering and/or Data Science
- Strong programming skills in Python
- Basic proficiency in PySpark and SQL
- Basic proficiency with Azure Databricks and cloud-based data engineering
- Experience in data and process modeling for large-scale systems
- Understanding Git and software engineering best practices
- Basic proficiency with data wrangling, transformation, and feature engineering
- Problem-solving skills and the ability to work independently
- Experience with MLOps and model deployment in production environments
- Conceptual understanding of NLP, AI/ML, and agentic networks
- Experience in implementing CI/CD pipelines for automated data workflows and model deployment, dockerization technologies
- Basic proficiency in Huggingface, Langchain and generative AI technologies for agentic networks
- Understanding of data streaming (e.g., Kafka, Azure Event Hubs)
- Knowledge of machine learning frameworks such as TensorFlow, PyTorch, or Scikit-Learn
Competitive Compensation & Growth Opportunities
- Dedicated training budget for conferences, online courses, and books to support continuous learning
- Access to English and Lithuanian language lessons
- Professional development through workshops, coaching sessions, and tech events
- Flexible working hours to suit your schedule
- Unlimited work-from-home option for greater autonomy
- Employee referral program with rewards up to 2000€ net
- Clients & External Ambassadors with rewards up to 5000€ net
- Social events, including Summer/Winter parties and a Dev Day celebration
- Team-building activities and annual live meet-ups with clients for enhanced collaboration
The final offer will depend on your experience and competencies
Key Skills
Ranked by relevanceReady to apply?
Join Helmes in Lithuania and take your career to the next level!
Application takes less than 5 minutes