Data is the center of my world. I enjoy that there are finally great places on the Internet get open large data sets such as Factual and Google Fusion Tables. However much of the data I want for different purposes is not neatly organized into an easily accessible data set or available on the Internet.
When I see data on the Internet that is controlled and not widely available I get the urge to set it free. Help the data owner organize it and make it available. Other times I need access to the data for business purposes.
This is when data harvesting or data scraping techniques come into play. It is when you programmatically pull a web page and parse the content to get at some or all of the data on the page. Then you setup some sort of script to spider and pull all of the available pages either through GET or POST of data to increment the site and encourage it to cough up all of its information.
There are many ways to pull these pages and crunch this data for whatever purposes. This isn’t what this post is about. It is more about ways to position these scrape or harvest scripts to grab the data.
See when I pull data from a known site, sometimes I making thousands, hundreds of thousands, or millions of requests for pages against a site. If the site owner is smart they will limit the number of requests you can make on their site from a single IP Address.
With limitations on the number of requests I can make in a given time period I have to slow my requests down. With large data sets this means I have to wait longer. Which just isn’t acceptable to my ADD personality. I can’t wait.
So I need a way of making requests from many IP addresses quickly.
Cloud Harvesting and Scraping Systems.
With the growth of cloud computing there are many places to quickly procure computing power at unique IP addresses. Using services such as Amazon EC2, Rackspace Cloud, and GoGrid makes it easy to quickly automate and distribute calls pulling remote pages across many IP addresses.
Of course this route costs $$. So to make it worth my time to deploy cloud infrastructure for harvesting and scraping I need a budget and a legitimate business purposes for the harvest.
Cloud Infrastructure as a Service (IaaS) is the quickest way to plan, deploy, maintain and shut down a distributed harvester and get data as fast you want. Of course you have to refrain from crossing to the dark side of cloud and becoming a Denial of Service (DOS) attack as well. How much can the target data source handle? How many requests can they deal with before their systems are overloaded and you start negatively impacting their business. This can be a fine line. With all the cloud computing power at your finger tips….you have to be careful.
Distributed Harvesting and Scraping System
Another way of creating a distributed approach to your harvesting and scraping program is to use a distributed approach that technically is still in the clouds, but not using major cloud computing providers. A common way is to procure accounts on many of the FREE hosting providers out their that utilize the language of your choice (PHP, Python) etc. You can easily setup a front site such as blog, forum, or other site and in the back ground be running scripts on timers that pull pages from your target host without alerting triggering their red flags. How you process the harvested data is up to you. I often send raw data back to central cloud based host for processing, cleanup and warehousing.
Both of these distributed solution create a sort of Ambient Computing Power that you have at your fingertips. You can then harvest data in a distributed fashion at whatever scale you have the budget for or time for depending on your business goals.
Harvesting of data can be fun if you are a data geek, but it can also be illegal. So be careful and wise in how you deploy.