Looking for a Freelance Data Quality Automation to join a team of rockstar developers. The candidate should have a minimum of 4+ yrs. of experience.There are multiple openings. If you're looking for freelance / part time opportunity (along with your day job) & a chance to work with the top 0.1% of developers in the industry, this one is for you! You will report into IIT'ans / BITS grads with 10+ years of development experience + work with F500 companies (our customers).Company Background - We are a multinational software company that is growing at a fast pace. We have offices in Florida & New Delhi. Our clientele spreads across the US, Australia & APAC. To give you a sense of our growth rate, we've added 70+ employees in the last 6 weeks itself and expect another 125+ by the end of Q4 2025.Key Responsibilities- Design, develop, and maintain automated test frameworks for data pipelines and ETL workflows.- Implement data validation and quality checks using Great Expectations and custom Python scripts.- Collaborate with Data Engineers to ensure data consistency, accuracy, and performance across Databricks and Snowflake environments.- Develop and execute PySpark-based test cases to validate data transformations and processing logic.- Integrate automated tests into CI / CD pipelines (e.g., GitHub Actions, Jenkins, Azure DevOps, or GitLab CI).- Monitor, troubleshoot, and improve data validation processes to ensure end-to-end data quality.- Work closely with cross-functional teams to define data testing strategies, validation metrics, and reporting dashboards.Required Skills & Qualifications- Proficient in Python (for scripting, data processing, and automation).- Hands-on experience with PySpark for large-scale data validation and transformation testing.- Strong understanding of Databricks workflows and Snowflake data warehousing.- Practical experience with Great Expectations for data quality checks.- Familiarity with CI / CD tools (GitHub Actions, Jenkins, GitLab, or Azure DevOps).- Good understanding of ETL processes, data pipelines, and data modeling.- Experience in Automation Testing and test-driven data development practices.- Excellent debugging, documentation, and communication skills.Preferred Skills- Experience with Airflow or other orchestration tools.- Exposure to AWS, Azure, or GCP data ecosystems.- Knowledge of SQL optimization and data performance tuning.- Familiarity with containerization tools (Docker, Kubernetes) is a plusWhat we need-~35 hours of work per week.-100% remote from our side-You will be paid out every month.-Min 4yrs of experience-Please apply only if you have a 100% remote job currently-If you do well, this will continue for a long time
Quality Automation • Delhi, Delhi, India