Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
We’re looking for a Data Engineer to join our team and help develop and maintain our modern data platform. In this role, you’ll design and build data pipelines while maintaining high standards of data quality and reliability, and collaborate closely with product and engineering teams to drive real business impact through data. Our data platform powers large-scale transformer inference that sits at the heart of our product.
The scope of the role is flexible and can grow with your experience. In addition to building data pipelines and workflows, there is room to influence the architecture of our data platform, technology choices, and how large-scale data and machine learning workloads run in production.
You’ll be part of the Data Engineering team, supported by Platform Engineering on cloud and infrastructure matters, so you can focus on building robust, scalable data systems.
What you’ll do:
- Develop and maintain our data platform, built on GCP and orchestrated with Dagster, and help evolve it for future needs
- Build and operate data workflows with strong observability, monitoring, data quality checks, and lineage
- Collaborate closely with cross-functional teams to deliver data solutions that drive product and business outcomes
- Work with open-source and modern data infrastructure technologies across orchestration, vector search, distributed processing and GPU workloads (e.g., Dagster, Vespa.ai, Ray Data, Kubernetes, GCP)
- Some experience with data engineering concepts such as data lakes, and modern data workflows
- Solid SQL skills and familiarity with Python and data infrastructure tools
- Python data stack experience (Polars, Arrow) - if you've wrangled large datasets in any context, scientific or otherwise, that counts
- Experience working in cloud or HPC environments (GCP is a plus) and familiarity with containers or job orchestration (Kubernetes, SLURM, or similar)
- Interest in open-source data tools and curiosity to learn how to select, operate, and maintain them
- A collaborative mindset and comfort working in a small team where you take ownership and solve practical problems
- You're eager to understand our domain and make engineering choices with business impact in mind
- Startup or scale-up experience is a plus.
- You can feel your impact in our small, driven company of around 50 people working globally
- Flexible hybrid work from Finland (Helsinki + remote) with focus on trust and results
- Six weeks of paid holiday + learning & development support + stock options
- A role where data and AI matter deeply, i.e. realistic problems, not hype
- A culture rooted in transparency, simplicity and reliability
We like to keep our process simple and transparent. Here’s what to expect:
- Conversation with Head of SW & Talent Acquisition Manager
- Conversation with Data Engineering Lead & Data Engineer/Chief Architect
- Final call with our CTO
Key Skills
Ranked by relevanceReady to apply?
Join IPRally and take your career to the next level!
Application takes less than 5 minutes

