Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
We´re looking for a 2 minimun Data Engineer Pyspark SQL AWS with fluent level of English.
- Minimum 3+ years of data engineering experience
- Minimum 2+ years of experience in Data Analytics in AWS.
- Experience in Apache Spark (PySpark)
- Experience in programming Python and SQL.
- Good knowledge on Orchestration of data pipelines (Airflow) and Containerization technologies (Docker and ECS).
LOCATION: 2 DAYS PER WEEK AT THE OFFICE / 3 DAYS REMOTE.
MADRID: CIUDAD LINEAL AREA.
MANDATORY: FLUENT LEVEL OF ENGLISH.
Key Skills
Ranked by relevanceReady to apply?
Join CAS Training and take your career to the next level!
Application takes less than 5 minutes

