Key Responsibilities :
- Design, develop, and maintain ETL pipelines and data workflows using Python and PySpark.
- Build and manage RESTful APIs for data access and integrations.
- Work closely with data scientists and analysts to deliver clean, reliable, and well-structured data.
- Optimize SQL queries, stored procedures, and performance for Oracle DB and MySQL.
- Implement and automate CI / CD pipelines for efficient deployment.
- Write unit and integration tests using PyTest.
- Parse and process XML files, handle file operations efficiently using Python.
- Ensure data quality through validation, cleansing, and monitoring pipelines in real-time.
- Follow Agile methodologies for iterative development and delivery.
- Apply Object-Oriented and Functional Programming principles in service design and architecture.
Required Skills :
Strong experience in Python (Backend only) with a focus on data structures , OOPs , and algorithms .Hands-on experience with PySpark for large-scale data processing.Solid knowledge of Django or Flask web frameworks.Expertise in SQL with a deep understanding of query tuning and stored procedures .Proficiency in file handling , XML parsing , and data validation using Python.Familiarity with REST API design and integration.Good knowledge of CI / CD tools like Git and Jenkins .Exposure to cloud platforms like GCP (preferred), AWS , or Azure .Excellent communication and interpersonal skills .Strong attention to detail and a problem-solving mindset .Good to Have Skills :
Certification in cloud (e.g., GCP, AWS) or Python-related technologies .Experience in working with financial domain applications or data-heavy systems .Skills Required
Aws, Azure, Python, Rest Api, Python Backend