Looking for a Freelance Data Quality Automation to join a team of rockstar developers. The candidate should have a minimum of 4+ yrs. of experience.
There are multiple openings. If you're looking for freelance / part time opportunity (along with your day job) & a chance to work with the top 0.1% of developers in the industry, this one is for you! You will report into IIT'ans / BITS grads with 10+ years of development experience + work with F500 companies (our customers).
Company Background - We are a multinational software company that is growing at a fast pace. We have offices in Florida & New Delhi. Our clientele spreads across the US, Australia & APAC. To give you a sense of our growth rate, we've added 70+ employees in the last 6 weeks itself and expect another 125+ by the end of Q4 2025.
Key Responsibilities
- Design, develop, and maintain automated test frameworks for data pipelines and ETL workflows.
- Implement data validation and quality checks using Great Expectations and custom Python scripts.
- Collaborate with Data Engineers to ensure data consistency, accuracy, and performance across Databricks and Snowflake environments.
- Develop and execute PySpark-based test cases to validate data transformations and processing logic.
- Integrate automated tests into CI / CD pipelines (e.g., GitHub Actions, Jenkins, Azure DevOps, or GitLab CI).
- Monitor, troubleshoot, and improve data validation processes to ensure end-to-end data quality.
- Work closely with cross-functional teams to define data testing strategies , validation metrics, and reporting dashboards.
Required Skills & Qualifications
Proficient in Python (for scripting, data processing, and automation).Hands-on experience with PySpark for large-scale data validation and transformation testing.Strong understanding of Databricks workflows and Snowflake data warehousing.Practical experience with Great Expectations for data quality checks.Familiarity with CI / CD tools (GitHub Actions, Jenkins, GitLab, or Azure DevOps).Good understanding of ETL processes , data pipelines, and data modeling.Experience in Automation Testing and test-driven data development practices.Excellent debugging, documentation, and communication skills.Preferred Skills
Experience with Airflow or other orchestration tools.Exposure to AWS, Azure, or GCP data ecosystems.Knowledge of SQL optimization and data performance tuning .Familiarity with containerization tools (Docker, Kubernetes) is a plusWhat we need
~35 hours of work per week.100% remote from our sideYou will be paid out every month.Min 4yrs of experiencePlease apply only if you have a 100% remote job currentlyIf you do well, this will continue for a long time