Skip to content

A scrapy pipeline which send items to Elastic Search server

Notifications You must be signed in to change notification settings

dameyerdave/scrapy-elasticsearch

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Description

Scrapy pipeline which allows you to store scrapy items in Elastic Search.

Install

pip install ScrapyElasticSearch

If you need support for ntlm:
pip install "ScrapyElasticSearch[extras]"

Usage (Configure settings.py:)

ITEM_PIPELINES = {
    'scrapyelasticsearch.scrapyelasticsearch.ElasticSearchPipeline': 500
}

ELASTICSEARCH_SERVERS = ['localhost']
ELASTICSEARCH_INDEX = 'scrapy'
ELASTICSEARCH_TYPE = 'items'
# Custom uniqe key
ELASTICSEARCH_UNIQ_KEY = 'url'

# Eather set a custom index split type and field
ELASTICSEARCH_INDEX_SUFFIX_TYPE = 'field'
ELASTICSEARCH_INDEX_SUFFIX_KEY = 'language'

# Or append a plain text
ELASTICSEARCH_INDEX_SUFFIX_TYPE = 'plain'
ELASTICSEARCH_INDEX_SUFFIX_KEY = 'mysuffix'

# Or split by date
ELASTICSEARCH_INDEX_DATE_FORMAT = '%Y-%m'

# can also accept a list of fields if need a composite key
ELASTICSEARCH_UNIQ_KEY = ['url', 'id']

ELASTICSEARCH_SERVERS - list of hosts or string (single host). Host format: protocl://username:password@host:port. Examples:

Available parameters (in settings.py:)

 ELASTICSEARCH_INDEX - elastic search index.
 ELASTICSEARCH_INDEX_SUFFIX_TYPE - the type of index splitting.
 ELASTICSEARCH_INDEX_SUFFIX_KEY - the key of the field to use the value for index splitting or the plain text you want to use depending on the type chosen.
 ELASTICSEARCH_INDEX_DATE_FORMAT - the format for date suffix for the index, see python datetime.strftime for format. Default is no date suffix.
 ELASTICSEARCH_TYPE - elastic search type
 ELASTICSEARCH_UNIQ_KEY - optional field, unique key in string (must be a field or a list declared in model, see items.py)
 ELASTICSEARCH_BUFFER_LENGTH - optional field, number of items to be processed during each bulk insertion to Elasticsearch. Default size is 500.
 ELASTICSEARCH_AUTH  - optional field, set to 'NTLM' to use NTLM authentification
 ELASTICSEARCH_USERNAME - optional field, set to 'DOMAIN\username', only used with NLTM authentification
 ELASTICSEARCH_PASSWORD - optional field, set to your 'password', only used with NLTM authentification

 ELASTICSEARCH_CA - optional settings to if es servers require custom CA files.
 Example:
 ELASTICSEARCH_CA = {
      'CA_CERT': '/path/to/cacert.pem',
      'CLIENT_CERT': '/path/to/client_cert.pem',
      'CLIENT_KEY': '/path/to/client_key.pem'
}

Here is an example app (dirbot https://github.com/jayzeng/dirbot) in case you are still confused.

Dependencies

See requirements.txt

Changelog

  • 0.9.1 Support different types to split the indices
  • 0.9: Accept custom CA cert to connect to es clusters
  • 0.8: Added support for NTLM authentification
  • 0.7.1: Added date format to the index name and a small bug fix
    • ELASTICSEARCH_BUFFER_LENGTH default was 9999, this has been changed to reflect documentation.
  • 0.7: A number of backwards incompatibility changes are introduced:
    • Changed ELASTICSEARCH_SERVER to ELASTICSEARCH_SERVERS
    • ELASTICSEARCH_SERVERS accepts string or list
    • ELASTICSEARCH_PORT is removed, you can specify it in the url
    • ELASTICSEARCH_USERNAME and ELASTICSEARCH_PASSWORD are removed. You can use this format ELASTICSEARCH_SERVERS=['http://username:password@host:port']
    • Changed scrapy.log to logging as scrapy now uses the logging module
  • 0.6.1: Able to pull configs from spiders (in addition to reading from config file)
  • 0.6: Bug fix
  • 0.5: Abilit to persist object; Option to specify logging level
  • 0.4: Remove debug
  • 0.3: Auth support
  • 0.2: Scrapy 0.18 support
  • 0.1: Initial release

Issues

If you find any bugs or have any questions, please report them to "issues" (https://github.com/knockrentals/scrapy-elasticsearch/issues)

Contributors

Licence

Copyright 2014 Michael Malocha

Expanded on the work by Julien Duponchelle

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

About

A scrapy pipeline which send items to Elastic Search server

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 100.0%