@@ -65,7 +65,7 @@ When you ran the command ``scrapy runspider quotes_spider.py``, Scrapy looked fo
65
65
Spider definition inside it and ran it through its crawler engine.
66
66
67
67
The crawl started by making requests to the URLs defined in the ``start_urls ``
68
- attribute (in this case, only the URL for quotes in *humor * category)
68
+ attribute (in this case, only the URL for quotes in the *humor * category)
69
69
and called the default callback method ``parse ``, passing the response object as
70
70
an argument. In the ``parse `` callback, we loop through the quote elements
71
71
using a CSS Selector, yield a Python dict with the extracted quote text and author,
@@ -83,9 +83,9 @@ While this enables you to do very fast crawls (sending multiple concurrent
83
83
requests at the same time, in a fault-tolerant way) Scrapy also gives you
84
84
control over the politeness of the crawl through :ref: `a few settings
85
85
<topics-settings-ref>`. You can do things like setting a download delay between
86
- each request, limiting amount of concurrent requests per domain or per IP, and
86
+ each request, limiting the amount of concurrent requests per domain or per IP, and
87
87
even :ref: `using an auto-throttling extension <topics-autothrottle >` that tries
88
- to figure out these automatically.
88
+ to figure these settings out automatically.
89
89
90
90
.. note ::
91
91
@@ -106,10 +106,10 @@ scraping easy and efficient, such as:
106
106
107
107
* Built-in support for :ref: `selecting and extracting <topics-selectors >` data
108
108
from HTML/XML sources using extended CSS selectors and XPath expressions,
109
- with helper methods to extract using regular expressions.
109
+ with helper methods for extraction using regular expressions.
110
110
111
111
* An :ref: `interactive shell console <topics-shell >` (IPython aware) for trying
112
- out the CSS and XPath expressions to scrape data, very useful when writing or
112
+ out the CSS and XPath expressions to scrape data, which is very useful when writing or
113
113
debugging your spiders.
114
114
115
115
* Built-in support for :ref: `generating feed exports <topics-feed-exports >` in
@@ -124,7 +124,7 @@ scraping easy and efficient, such as:
124
124
well-defined API (middlewares, :ref: `extensions <topics-extensions >`, and
125
125
:ref: `pipelines <topics-item-pipeline >`).
126
126
127
- * Wide range of built-in extensions and middlewares for handling:
127
+ * A wide range of built-in extensions and middlewares for handling:
128
128
129
129
- cookies and session handling
130
130
- HTTP features like compression, authentication, caching
0 commit comments