From Wikipedia, the free encyclopedia
  (Redirected from ScraperWiki)
Jump to navigation Jump to search
The ScraperWiki logo, a wheel tractor-scraper.
Available inEnglish
RevenueSponsored by 4iP[1]
Alexa rankPositive decrease 133,089 (April 2014)[2]
Current statusActive
Content license
Affero General Public License[3]

QuickCode (formerly ScraperWiki) is a web-based platform for collaboratively building programs to extract and analyze public (online) data, in a wiki-like fashion. "Scraper" refers to screen scrapers, programs that extract data from websites. "Wiki" means that any user with programming experience can create or edit such programs for extracting new data, or for analyzing existing datasets.[1] The main use of the website is providing a place for programmers and journalists to collaborate on analyzing public data.[4][5][6][7][8][9]

The service was renamed circa 2016, as "it isn't a wiki or just for scraping any more".[10] At the same time, the eponymous parent company was renamed 'The Sensible Code Company'.[10]


Scrapers are created using a browser based IDE or by connecting via SSH to a server running Linux. They can be programmed using a variety of programming languages, including Perl, Python, Ruby, JavaScript and R.


ScraperWiki was founded in 2009 by Julian Todd and Aidan McGuire. It was initially funded by 4iP, the venture capital arm of TV station Channel 4. Since then, it has attracted an additional £1 Million round of funding from Enterprise Ventures.

Francis Irving is the chief executive officer of ScraperWiki.[11]

See also[edit]


  1. ^ a b Jamie Arnold (2009-12-01). "4iP invests in ScraperWiki". 4iP.
  2. ^ "Scraperwiki.com Site Info". Alexa Internet. Retrieved 2014-04-01.
  3. ^ "GNU Affero General Public License v3.0 - sensiblecodeio". GitHub. Retrieved 30 December 2017.
  4. ^ Cian Ginty (2010-11-19). "Hacks and hackers unite to get solid stories from difficult data". The Irish Times.
  5. ^ Paul Bradshaw (2010-07-07). "An introduction to data scraping with Scraperwiki". Online Journalism Blog.
  6. ^ Charles Arthur (2010-11-22). "Analysing data is the future for journalists, says Tim Berners-Lee". The Guardian.
  7. ^ Deirdre McArdle (2010-11-19). "In The Papers 19 November". ENN.
  8. ^ "Journalists and developers join forces for Lichfield 'hack day'". The Lichfield Blog. 2010-11-15.
  9. ^ Alison Spillane (2010-11-17). "Online tool helps to create greater public data transparency". Politico.
  10. ^ a b "ScraperWiki". ScraperWiki. Retrieved 7 February 2017.
  11. ^ Francis Irving, Rufus Pollock (9 March 2012). "From CMS to DMS: C is for Content, D is for Data". Open Knowledge Blog.CS1 maint: Uses authors parameter (link)

External links[edit]