赞
踩
pip install scrapy_redis
原来代码:
class XiaohuaSpider(scrapy.Spider):
pass
修改后为:
from scrapy_redis.spiders import RedisSpider
class XiaohuaSpider(RedisSpider):
pass
# start_urls = ['http://www.xiaohuar.com/list-1-0.html']
redis_key = "xiaohua1"
custom_settings = { "ROBOTSTXT_OBEY": False, "CONCURRENT_REQUESTS": 16, "DOWNLOAD_DELAY": 1, "COOKIES_ENABLED": False, "ITEM_PIPELINES": { # Store scraped item in redis for post-processing. 分布式redispipeline 'scrapy_redis.pipelines.RedisPipeline': 299, # 将爬取的数据存储到mysql中,定义管道文件 'xiaohuaSqider.pipelines.MySQLTwistedPipeline': 300, }, """ scrapy-redis配置 """ # Enables scheduling storing requests queue in redis. "SCHEDULER": "scrapy_redis.scheduler.Scheduler", # Ensure all spiders share same duplicates filter through redis. "DUPEFILTER_CLASS": "scrapy_redis.dupefilter.RFPDupeFilter", "REDIS_HOST": "192.168.4.193", # 本机ip "REDIS_PORT": "6379" # redis端口号,默认6379 }
grant all privileges on *.* to '用户名'@'%' identified by '密码';
flush privileges;
redis-cli.exe -h 192.168.4.193 -p 6379
lpush redis_key的值 网址
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。