Data Engineer - PySpark - Airflow - Unix - Docker/Kubernetes - CI/CD - Finland

Uusimaa  ‐ Onsite
This project has been archived and is not accepting more applications.
Browse open projects on our job board.

Keywords

Pyspark Airflow Unix Continuous Integration Kubernetes Docker Github Business To Business Batch Processing Extract Transform Load (ETL) Data Quality Governance Maintenance Data Streaming Data Pipeline

Description

Data Engineer - PySpark - Airflow - Unix - Docker/Kubernetes - CI/CD - Finland

Empiric has received an exciting opportunity for an Data Engineer with good experience in PySpark, Airflow and proficiency in Unix. Docker and Kubernetes experience are also required.

The Data Engineer will troubleshoot data pipelines and address issues with Real Time, batch processing.

Plus Develop ETL process and data flows using PySpark and Airflow. Maintain and troubleshoot CI/CD pipeline issues using GitHub Actions and JFrog

Hybrid 6 to 10 days in the clients Helsinki office initially. Can be reduced in time

Skills/Experience:

  • PySpark & Airlow
  • Unix
  • Docker & Kubernetes
  • CI/CD Tools: GitHub Actions
  • Maintenance & troubleshooting
  • Understanding & ensuring data quality, integrity, governance
  • Term: 6 to 18 Months plus extensions
  • Good day rate (B2B is fine) + Starter Bonus + Free Lunch Club experience

This is a critical position, please respond to this advert or reach out to Woody on either (see below) or (whatsapp is fine) for a confidential chat and more details on rate and this terrific project.

Start date
July/Aug
Duration
1 to 2 years + extensions
(extension possible)
From
Empiric Solutions
Published at
02.07.2025
Project ID:
2893876
Industry
IT
Contract type
Freelance
To apply to this project you must log in.
Register