ScraperWiki

  • ScraperWiki: brief description

    ScraperWiki is a hosted environment for writing automated processes to scan public websites and extract structured information from the pages they’ve published. It handles all of the boilerplate code that you normally have to write to handle crawling websites, gives you a simple online editor for your Ruby, Python, or PHP scripts, and automatically runs your crawler as a background process.