Job Title: Associate Developer (Web Scraping and Data Extraction Specialist)
Relevant experience: 1+ years
Location: Mumbai
About Us:
Join the Ad Tech Disruptors! We're India's hottest marketplace Ad Tech data company, and we
are revolutionizing how Amazon and Flipkart sellers manage their advertising data and optimize
across platforms. Our flagship product, Hector, is the most sophisticated data visualization,
analytics, and optimization platform for Amazon & Flipkart sellers globally. Built by performance
marketing specialists, Hector empowers brands to scale profitably by presenting data in an
optimized manner for granular analysis and strategic optimization decisions.
Role Overview:
We are looking for an experienced Python Developer with strong experience in web scraping
and data extraction using tools like Selenium, Beautiful Soup etc.
Key Responsibilities:
Develop and maintain web crawlers using Python, Selenium, and Beautiful Soup to
extract data from various sources.
Write efficient, scalable, and reusable code to handle large volumes of data.
Identify and resolve any issues or bottlenecks in the web scraping process.
Collaborate with the data engineering team to ensure seamless integration of extracted
data into the data pipeline.
Monitor and maintain the web scraping infrastructure to ensure reliability and accuracy of
data collection.
Keep up-to-date with the latest developments in web scraping technologies and best
practices.
Requirements:
Proven experience as a Python Developer, specifically with web scraping and data
extraction.
Proficiency in using Selenium for browser automation.
Strong knowledge of Beautiful Soup or similar tool for HTML parsing and data extraction.
Experience with other web scraping frameworks and libraries is a plus.
Familiarity with handling and parsing JSON and XML data.
Strong problem-solving skills and attention to detail.
Ability to work independently and in a team-oriented environment.
Good communication and documentation skills.
Preferred Qualifications:
Experience with cloud platforms (e.g., AWS, GCP) for deploying and managing web
crawlers & data.
Knowledge of data storage solutions such as SQL, Big Query and No SQL databases.