Skip to content

Commit

Permalink
minor corrections in documentation.
Browse files Browse the repository at this point in the history
  • Loading branch information
bagratte committed Apr 19, 2015
1 parent beea926 commit 1312bcd
Show file tree
Hide file tree
Showing 4 changed files with 9 additions and 9 deletions.
2 changes: 1 addition & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -12,4 +12,4 @@ dist
.idea

# Windows
Thumbs.db
Thumbs.db
4 changes: 2 additions & 2 deletions docs/topics/feed-exports.rst
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Feed exports

One of the most frequently required features when implementing scrapers is
being able to store the scraped data properly and, quite often, that means
generating a "export file" with the scraped data (commonly called "export
generating an "export file" with the scraped data (commonly called "export
feed") to be consumed by other systems.

Scrapy provides this functionality out of the box with the Feed Exports, which
Expand All @@ -21,7 +21,7 @@ Serialization formats
=====================

For serializing the scraped data, the feed exports use the :ref:`Item exporters
<topics-exporters>` and these formats are supported out of the box:
<topics-exporters>`. These formats are supported out of the box:

* :ref:`topics-feed-format-json`
* :ref:`topics-feed-format-jsonlines`
Expand Down
10 changes: 5 additions & 5 deletions docs/topics/item-pipeline.rst
Original file line number Diff line number Diff line change
Expand Up @@ -5,14 +5,14 @@ Item Pipeline
=============

After an item has been scraped by a spider, it is sent to the Item Pipeline
which process it through several components that are executed sequentially.
which processes it through several components that are executed sequentially.

Each item pipeline component (sometimes referred as just "Item Pipeline") is a
Python class that implements a simple method. They receive an item and perform
an action over it, also deciding if the item should continue through the
pipeline or be dropped and no longer processed.

Typical use for item pipelines are:
Typical uses of item pipelines are:

* cleansing HTML data
* validating scraped data (checking that the items contain certain fields)
Expand Down Expand Up @@ -167,7 +167,7 @@ Duplicates filter
-----------------

A filter that looks for duplicate items, and drops those items that were
already processed. Let say that our items have an unique id, but our spider
already processed. Let's say that our items have a unique id, but our spider
returns multiples items with the same id::


Expand Down Expand Up @@ -198,6 +198,6 @@ To activate an Item Pipeline component you must add its class to the
}

The integer values you assign to classes in this setting determine the
order they run in- items go through pipelines from order number low to
high. It's customary to define these numbers in the 0-1000 range.
order in which they run: items go through from lower valued to higher
valued classes. It's customary to define these numbers in the 0-1000 range.

2 changes: 1 addition & 1 deletion docs/topics/link-extractors.rst
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@ LxmlLinkExtractor
module.
:type deny_extensions: list

:param restrict_xpaths: is a XPath (or list of XPath's) which defines
:param restrict_xpaths: is an XPath (or list of XPath's) which defines
regions inside the response where links should be extracted from.
If given, only the text selected by those XPath will be scanned for
links. See examples below.
Expand Down

0 comments on commit 1312bcd

Please sign in to comment.