我有一只蜘蛛可以下载特定网站的jpg。在过去,我已经在图片管道中解析了response.url
,以便在文件被下载时重命名该文件。问题是网站的目录结构很奇怪,因此解析image_urls
以重命名目标文件并不起作用。作为一种解决方法,我只使用原始图形名称作为文件。
我想使用来自实际Scrapy对象本身的数据,但我似乎无法将变量传递到蜘蛛的图像管道中。从下面的代码中,我想在蜘蛛中解析url
并将其作为变量传递给管道中的otImagesPipeline
,但没有任何作用。我尝试查看Scrapy文档,但无法找到如何执行此操作。
这可以用Scrapy吗?
这是我的蜘蛛代码:
settings.py:
BOT_NAME = 'bid'
MEDIA_ALLOW_REDIRECTS = True
SPIDER_MODULES = ['bid.spiders']
NEWSPIDER_MODULE = 'bid.spiders'
ITEM_PIPELINES = {'bid.pipelines.otImagesPipeline': 1}
IMAGES_STORE = 'C:\\temp\\images\\filenametest'
pipelines.py
import scrapy
from scrapy.contrib.pipeline.images import ImagesPipeline
class otImagesPipeline(ImagesPipeline):
def file_path(self, request, response=None, info=None):
targetfile = request.url.split('/')[-1]
return targetfile
items.py
import scrapy
class BidItem(scrapy.Item):
url = scrapy.Field()
title = scrapy.Field()
caption = scrapy.Field()
image_urls = scrapy.Field()
getbid.py (蜘蛛)
import scrapy
from scrapy.linkextractors import LinkExtractor
from scrapy.spiders import CrawlSpider, Rule
from bid.items import BidItem
from urllib import parse as urlparse
class GetbidSpider(CrawlSpider):
name = 'getbid'
allowed_domains = ['example.com']
start_urls = ['http://www.example.com']
rules = (
Rule(LinkExtractor(), callback='parse_item', follow=True),
)
def parse_item(self, response):
for sel in response.xpath('//a'):
link = str(sel.xpath('@href').extract()[0])
if (link.endswith('.jpg')):
href = BidItem()
href['url'] = response.url
href['title'] = response.css("h1.entry-title::text").extract_first()
href['caption'] = response.css("p.wp-caption-text::text").extract()
href['image_urls'] = [link]
yield href
yield scrapy.Request(urlparse.urljoin('http://www.example.com/',link),callback=self.parse_item)
更新
感谢Umair的帮助,我能够完全按照我的需要修复它。 以下是修订后的代码:
getbid.py
def parse_item(self, response):
for sel in response.xpath('//a'):
link = str(sel.xpath('@href').extract()[0])
if (link.endswith('.jpg')):
href = BidItem()
href['url'] = response.url
href['title'] = response.css("h1.entry-title::text").extract_first()
href['caption'] = response.css("p.wp-caption-text::text").extract()
future_dir = href['url'].split("/")[-2]
href['images'] = {link: future_dir}
yield href
yield scrapy.Request(urlparse.urljoin(http://www.example.com/',link),callback=self.parse_item)
pipelines.py
class otImagesPipeline(ImagesPipeline):
def get_media_requests(self, item, info):
if 'images' in item:
for image_url, img_dir in item['images'].items():
request = scrapy.Request(url=image_url)
request.meta['img_dir'] = img_dir
yield request
def file_path(self, request, response=None, info=None):
filename = request.url.split('/')[-1]
filedir = request.meta['img_dir']
filepath = filedir + "/" + filename
return filepath
答案 0 :(得分:2)
您的Spider课程中有IMAGES_STORE
,因此您可以稍后在ImagesPipeline
file_path
方法
class GetbidSpider(CrawlSpider):
name = 'getbid'
IMAGE_DIR = 'C:\\temp\\images\\filenametest'
custom_settings = {
"IMAGES_STORE": IMAGE_DIR
}
allowed_domains = ['example.com']
start_urls = ['http://www.example.com']
rules = (
Rule(LinkExtractor(), callback='parse_item', follow=True),
)
def parse_item(self, response):
for sel in response.xpath('//a'):
link = str(sel.xpath('@href').extract()[0])
if (link.endswith('.jpg')):
href = BidItem()
href['url'] = response.url
href['title'] = response.css("h1.entry-title::text").extract_first()
href['caption'] = response.css("p.wp-caption-text::text").extract()
href['images'] = {link: href['title']}
yield href
yield scrapy.Request(urlparse.urljoin('http://www.example.com/',link),callback=self.parse_item)
然后在ImagesPipeline
class CustomImagePipeline(ImagesPipeline):
def get_media_requests(self, item, info):
if 'images' in item:
for image_url, img_name in item['images'].iteritems():
request = scrapy.Request(url=image_url)
request.meta['img_name'] = img_name
yield request
def file_path(self, request, response=None, info=None):
return os.path.join(info.spider.IMAGE_DIR, request.meta['img_name'])