Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
SUMMARY
馃挵 175 - 200 PLN/h on B2B
馃挵 20 000 - 24 000 PLN gross on UoP
馃搷 Remote
馃捈 Senior
PROJECT
We are looking for an experienced Data Scientist to support the development and continuous enhancement of a large-scale data and machine learning ecosystem used in next-generation automotive solutions.
The project involves building Spark-based data pipelines, improving data quality processes, implementing model evaluation workflows, and developing robust monitoring for ML models running in production.
The role includes both initial development of the platform components and long-term maintenance and optimization.
YOU WILL
Initial Development:
- Develop Spark jobs for data ingestion and feature engineering
- Implement data quality monitoring (metrics, dashboards, alerting)
- Build logic for model evaluation and automated deployment decisions
- Develop model monitoring with visualized KPIs and technical metrics
Further Development / Maintenance:
- Continuously extend data pipelines and feature engineering workflows
- Enhance data quality metrics and monitoring coverage
- Expand model monitoring logic and dashboards
- Troubleshoot and fix code issues, including edge cases
- Experiment with new ML algorithms and additional data attributes
- Optimize performance and cost (algorithms, data structures, storage formats)
- Adjust training/deployment pipeline configurations (frequency, resources, etc.)
MUST HAVE
- 5+ years of experience
- Strong commercial experience with PySpark
- Excellent knowledge of Python
- Practical experience with GitHub
- Strong data analysis skills (Jupyter, Seaborn, exploratory analytics)
- Solid SQL knowledge
- Experience with Kubeflow or MLflow (MLOps frameworks for training, deployment & monitoring)
- Understanding of MLOps practices, including continuous training
- Experience with ML frameworks: scikit-learn, Pandas, Optuna
- Ability to create Grafana dashboards
- General knowledge of AWS services (S3, IAM, etc.)
- In-depth understanding of statistics and machine learning (missing data, outliers, model validation, algorithms)
- Fluent in Polish and good English
NICE TO HAVE
- Experience optimizing data pipelines (Iceberg, Parquet, DynamoDB, etc.)
- Background in automotive or IoT data projects
- Experience with cost optimization for ML systems
- Experience with large-scale model deployment pipelines
OUR OFFER
- Relocation package (4500 PLN total value), paid in three installments (1500 PLN per month)
- Benefits: Extended medical care (over 2000 medical facilities in Poland, 80 in Krak贸w) for you and your family; Multisport Benefit card; Life insurance
- Remote work option
Key Skills
Ranked by relevanceReady to apply?
Join Infolet and take your career to the next level!
Application takes less than 5 minutes

