From Wikipedia, the free encyclopedia
Jump to navigation Jump to search
Scrapy logo.jpg
Developer(s)Zyte (formerly Scrapinghub)
Initial release26 June 2008 (2008-06-26)
Stable release
2.6.2[1] Edit this on Wikidata / 25 July 2022; 2 months ago (25 July 2022)
Written inPython
Operating systemWindows, macOS, Linux
TypeWeb crawler
LicenseBSD License Edit this on Wikidata

Scrapy (/ˈskrp/[2] SKRAY-peye) is a free and open-source web-crawling framework written in Python and developed in Cambuslang. Originally designed for web scraping, it can also be used to extract data using APIs or as a general-purpose web crawler.[3] It is currently maintained by Zyte (formerly Scrapinghub), a web-scraping development and services company.

Scrapy project architecture is built around "spiders", which are self-contained crawlers that are given a set of instructions. Following the spirit of other don't repeat yourself frameworks, such as Django,[4] it makes it easier to build and scale large crawling projects by allowing developers to reuse their code.

The Scrapy framework provides you with powerful features such as auto-throttle, rotating proxies and user-agents, allowing you scrape virtually undetected across the net. Scrapy also provides a web-crawling shell, which can be used by developers to test their assumptions on a site’s behavior.[5]

Some well-known companies and products using Scrapy are: Lyst,[6][7],[8] Sayone Technologies,[9] Sciences Po Medialab,[10]’s World Government Data site.[11][1]


Scrapy was born at London-based web-aggregation and e-commerce company Mydeco, where it was developed and maintained by employees of Mydeco and Insophia (a web-consulting company based in Montevideo, Uruguay). The first public release was in August 2008 under the BSD license, with a milestone 1.0 release happening in June 2015.[12] In 2011, Zyte (formerly Scrapinghub) became the new official maintainer.[13][14]


  1. ^; publication date: 25 July 2022; retrieved: 3 September 2022.
  2. ^ Commit 975f150
  3. ^ Scrapy at a glance.
  4. ^ "Frequently Asked Questions". Retrieved 28 July 2015.
  5. ^ "Scrapy shell". Retrieved 28 July 2015.
  6. ^ Bell, Eddie; Heusser, Jonathan. "Scalable Scraping Using Machine Learning". Archived from the original on 4 June 2016. Retrieved 28 July 2015.
  7. ^ Scrapy | Companies using Scrapy
  8. ^ Montalenti, Andrew. "Web Crawling & Metadata Extraction in Python".
  9. ^ "Scrapy Companies". Scrapy website.
  10. ^ Hyphe v0.0.0: the first release of our new webcrawler is out!
  11. ^ Ben Firshman [@bfirsh] (21 January 2010). "World Govt Data site uses Django, Solr, Haystack, Scrapy and other exciting buzzwords #opendata #datastore" (Tweet) – via Twitter.
  12. ^ Medina, Julia (19 June 2015). "Scrapy 1.0 official release out!". scrapy-users (Mailing list).
  13. ^ Pablo Hoffman (2013). List of the primary authors & contributors. Retrieved 18 November 2013.
  14. ^ Interview Scraping Hub.

External links[edit]