- Scrape Yell.com and save data into CSV (Business Name, Phone No, Address, Website)
- Check if Website is up and remove duplicate websites before continuing
- Scrape Email from Landing Page of Website using BeautifulSoup (Not works for emails generated from JS)
- Scrape Emails from all pages of a Website using Selenium (to counter above problem)
- Add Thomson Local support
- Create a Separate Scraper for Facebook business pages
- Integrate facebook scraper with extract_all_emails.py
- Create a check for finding facebook links with emails when scraping web pages and use the facebook links to find emails
- Integrate All Modules
- Use Proxy Servers in Python to avoid getting blacklisted
- Implement timeout based retry when scraping websites
- Create Scraper for other pages mentioned below
- Create a tkinter based UI
- Thompsons directory (Easy and Contain Emails + Websites)
- Check a trade - Contains no emails (Scraper should Scrape name, website, phone number)
- (Abandoned) Rated people - (Cannot custom search, need to select from available options)
- (Requires Proxy) My builder