We are looking for a skilled Data Engineer with strong expertise in PySpark, Databricks, and modern data engineering practices.
Requirements
- Design, develop, and maintain scalable ETL/ELT data pipelines
- Work extensively with PySpark and Apache Spark for large-scale data processing
- Build and manage workflows using Apache Airflow
- Develop and optimize data solutions on Databricks (Jobs, Delta Lake)
- Work with cloud-based data lakes (S3 or equivalent)
- Write efficient and complex SQL queries for data transformation and analysis
- Run and manage Spark workloads on EMR Serverless or other managed Spark platforms
- Ensure data quality, reliability, and performance optimization of pipelines
To apply for this job please visit job-boards.greenhouse.io.

Follow us on social media