About Client : -
Our client is a French multinational information technology (IT) services and consulting company, headquartered in Paris, France. Founded in 1967, It has been a leader in business transformation for over 50 years, leveraging technology to address a wide range of business needs, from strategy and design to managing operations.
The company is committed to unleashing human energy through technology for an inclusive and sustainable future, helping organisations accelerate their transition to a digital and sustainable world.
They provide a variety of services, including consulting, technology, professional, and outsourcing services.
Job Details : -
location : Pune
Mode Of Work : Hybrid
Notice Period : Immediate Joiners
Experience : 8-10 yrs
Type Of Hire : Contract to Hire
JOB DESCRIPTION :
Design, develop, and maintain scalable data pipelines using PySpark.
Work with large-scale datasets in distributed computing environments.
Implement ETL / ELT processes to ingest, transform, and load data from various sources.
Deploy and manage data workflows on AWS services such as S3, EMR, Glue, Lambda, Redshift, Athena, etc.
Collaborate with data scientists, analysts, and other engineers to understand data requirements.
Monitor and optimize performance of data pipelines and cloud infrastructure.
Ensure data quality, integrity, and security across all stages of the pipeline.
Required Skills :
Strong programming skills in Python with hands-on experience in PySpark.
Proficiency in AWS cloud services related to data engineering.
Experience with data lake and data warehouse architectures.
Familiarity with SQL and NoSQL databases.
Knowledge of CI / CD pipelines, version control (e.g., Git), and containerization (e.g., Docker) is a plus.
Excellent problem-solving and communication skills.
Pyspark Developer • Guntur, IN