Note : It's a Data Scrapping / Web Scrapping based role for Bangalore & Gurugram location.
Mandatory Skills : Python and any scrapping framework (Scrapy, Selenium, Playwright, or BeautifulSoup).
Role : Data Acquisition Engineer
Location : Bangalore / Gurgaon
Experience : 3-5 years
An ideal candidate has minimum 3 years of hands-on programming experience in Python, HTML and JavaScript. Candidate has demonstrable knowledge in web development or web scraping and is comfortable working with JSON / XML files to extract data.
As a Data Engineer focused on web crawling and platform data acquisition,
- You will design, develop, and maintain large-scale web scraping pipelines to extract valuable platform data.
- You will be responsible for implementing scalable and resilient data extraction solutions, ensuring seamless data retrieval while working with proxy management, anti-bot bypass techniques, and data parsing.
- Optimizing scraping workflows for performance, reliability, and efficiency will be a key part of your role.
- Additionally, you will ensure that all extracted data maintains high quality and integrity.
Our expectations from the right candidate :
Strong experience in Python and web scraping frameworks such as Scrapy, Selenium, Playwright, or BeautifulSoup.Knowledge of distributed web crawling architectures and job scheduling.Familiarity with headless browsers, CAPTCHA-solving techniques, and proxy management to handle dynamic web challenges.Experience with data storage solutions, including SQL, and cloud storage.Understanding of big data technologies like Spark and Kafka (a plus).Strong debugging skills to adapt to website structure changes and blockers.