Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
We’re looking for a Data Engineer to join our team and play a key role in developing and maintaining our modern data platform. In this role, you’ll design, build, and own data pipelines, ensure high standards of data quality and reliability, and collaborate closely with product and engineering teams to drive real business impact through data. Our core data workflow performs large-scale inference with a transformer model at the heart of our product.
You’ll be part of the Data Engineering team, supported by Platform Engineering on cloud and infrastructure matters, so you can focus on building robust, scalable data systems.
What You’ll Do
- Develop and maintain our data platform, built on GCP and orchestrated with Dagster, and help it evolve it for future needs
- Implement ELT pipelines and data workflows; optimise for performance, observability and reliability
- Build monitoring, data quality checks and lineage visibility so our datasets are trustworthy
- Collaborate closely with cross-functional teams to deliver data solutions that drive product and business outcomes
- Work with open-source tooling, host and manage infrastructure
- Leverage technologies like Dagster, Vespa.ai, GCP, PostgreSQL, Ray Data, Kubernetes and GPUs
- Proven experience in data engineering: data lakes / lakehouse architectures, ELT pipelines and modern data workflows
- Strong SQL skills and experience with both Python and data infrastructure.
- Familiarity with cloud environments (GCP preferred) and container/orchestration tools (like Kubernetes)
- You’ve worked with open-source data infrastructure and know how to choose, run and maintain tools
- You are a doer, collaborative, and ready to tackle challenging problems with good judgment
- You are self-driven and thrive in a smaller team environment where your contributions are visible and influence both your own work and that of your colleagues
- You're eager to understand our domain and make engineering choices with business impact in mind
- Startup or scale-up experience is a plus (you’re comfortable with visible details and less dependency on larger teams)
- You can feel your impact in our small, driven company of around 50 people working globally
- Flexible hybrid work from Finland (Helsinki + remote) with focus on trust and results
- Six weeks of paid holiday + learning & development support + stock options
- A role where data and AI matter deeply, i.e. realistic problems, not hype
- A culture rooted in transparency, simplicity and reliability
We like to keep our process simple and transparent. Here’s what to expect:
- Introductory call with Talent Acquisition Manager
- Conversation with Head of SW & Lead Data Engineer
- Conversation with ML Infrastructure Lead & Data Engineer/Chief Architect
- Final call with our CTO
Key Skills
Ranked by relevanceReady to apply?
Join IPRally and take your career to the next level!
Application takes less than 5 minutes

