Skip to main content

ScraperWiki Turns Web Pages Into Usable Data

posted onMarch 20, 2011
by hitbsecnews

A scraper is a program written to take content off of a webpage or other data source and turn it into some kind of usable format, usually an RSS feed or by entering it directly into a database. Designing a scraper can be tricky as each site is different, ScraperWiki aims to fix this by creating a repository of these scripts with a goal to ease the pain of designing them.

An example use of a scraper: let's say a government entity releases daily information regarding finances, and you want to graph or otherwise track this data for personal or business use. Going to the website each day and entering the data manually is certainly one labor-intensive way to do it, but as with any good hacker will tell you - if you have to do anything more than once it is better to automate it.

ScraperWiki is a centralized location for these custom built scrapers. Instead of writing your own from scratch, you can search their database to see if a scraper has already been written for a source. One of the functions of ScraperWiki is to support open government initiatives. The Big Clean is actually being held today with the goal of opening local government data with the help of scrapers and data processors.

Source

Tags

Software-Programming

You May Also Like

Recent News

Friday, November 29th

Tuesday, November 19th

Friday, November 8th

Friday, November 1st

Tuesday, July 9th

Wednesday, July 3rd

Friday, June 28th

Thursday, June 27th

Thursday, June 13th

Wednesday, June 12th

Tuesday, June 11th