Job Summary :
We are seeking an experienced Senior Python Developer with deep expertise in Google Cloud Platform (GCP), PySpark, and Apache Airflow. The ideal candidate will be responsible for designing, developing, and maintaining scalable data pipelines, orchestrating complex workflows using Airflow, and optimizing data processing jobs in a distributed environment.
Key Responsibilities :
- Design and develop robust, scalable, and efficient data pipelines using Python and PySpark.
- Create, schedule, and monitor DAGs in Apache Airflow for data workflow orchestration.
- Work with GCP services (such as BigQuery, Cloud Composer, GCS, Pub / Sub, Dataflow, etc.) to build end-to-end data solutions.
- Implement data transformation, cleansing, aggregation, and validation logic in pipelines.
- Collaborate with Data Engineers, Analysts, and other stakeholders to understand business requirements and translate them into technical solutions.
- Optimize performance and cost-efficiency of data pipelines on GCP.
- Monitor production workflows and troubleshoot issues to ensure data accuracy and availability.
- Follow coding best practices and maintain high standards for data quality, security, and compliance.
- Participate in code reviews, testing, and deployment processes in CI / CD environments.
Required Skills :
7+ years of professional experience in software development with a strong focus on Python.3+ years hands-on experience with Apache Airflow for orchestration and automation.3+ years of experience working with PySpark and large-scale data processing.Strong hands-on experience with Google Cloud Platform (GCP), especially services like :BigQueryCloud Composer (Airflow)Cloud Storage (GCS)Pub / SubDataflow (optional)Solid understanding of data modeling, ETL / ELT pipelines, and batch / stream processing.Proficient in SQL and working with large datasets.Experience with version control (Git) and CI / CD pipelines.Familiarity with containerization (Docker) and Kubernetes is a plus.Preferred Qualifications :
GCP Certification (e.g., Professional Data Engineer or Cloud Developer) is a plus.Experience with Terraform or other infrastructure-as-code tools.Exposure to other workflow tools (e.g., Luigi, Prefect) is a plus.Strong problem-solving and analytical skills.Excellent communication and collaboration skills.(ref : hirist.tech)