Job Title : PySpark Developers
Experience Level : 6 to 9 years
Location : Hisar (Haryana), Gurgaon (Preferable), NCR region
Job Responsibilities
- Develop and maintain scalable data pipelines using Python and PySpark.
- Collaborate with data engineers and data scientists to understand and fulfill data processing needs.
- Optimize and troubleshoot existing PySpark applications for performance improvements.
- Write clean, efficient, and well-documented code following best practices.
- Participate in design and code reviews.
- Develop and implement ETL processes to extract, transform, and load data.
- Ensure data integrity and quality throughout the data lifecycle.
- Stay current with the latest industry trends and technologies in big data and cloud computing.
Skill Set and Experience
Minimum 6 years of experience in design and implementation of systems using PySpark, database migration, transformation and integration solutions for any Data warehousing project.Experience in Apache Spark and Python programming experienceExperience in developing data processing tasks using PySpark such as reading data from external sources, merging data, performing data enrichment and loading into target data destinations.Experience with building APIs for provisioning data to downstream systems by leveraging different frameworks.Hands-on experience on Jupyter Notebook / Zeppelin / PyCharm etc. IDEs.Preferred Skills
PythonPySparkBig DataDistributed ComputingETL ProcessesData EngineeringCloud Computing (Azure preferred)Data WarehousingApache Spark