Data Engineer (Databrick + Pyspark)

Hybrid Full TimePune, Maharashtra, IndiaCapco

We are looking for a skilled Data Engineer with strong expertise in PySpark, Databricks, and modern data engineering practices.

Requirements

  • Design, develop, and maintain scalable ETL/ELT data pipelines
  • Work extensively with PySpark and Apache Spark for large-scale data processing
  • Build and manage workflows using Apache Airflow
  • Develop and optimize data solutions on Databricks (Jobs, Delta Lake)
  • Work with cloud-based data lakes (S3 or equivalent)
  • Write efficient and complex SQL queries for data transformation and analysis
  • Run and manage Spark workloads on EMR Serverless or other managed Spark platforms
  • Ensure data quality, reliability, and performance optimization of pipelines

To apply for this job please visit job-boards.greenhouse.io.


You can apply to this job and others using your online resume. Click the link below to submit your online resume and email your application to this employer.

Tired of manual job applications?

JobCopilot auto-applies to thousands of RevOps and GTM roles on your behalf — so you can focus on interviews, not applications.

Applying for this role?

Tailor your resume to this exact role — hiring managers notice the difference.

Latest articles on the blog

RECRUITERS!

Reduce the risk of your recruitment process (applicant quality, long and inefficient process) by selecting from a relevant pool of candidates.

POST A NEW JOB NOW!