嘗試使用scrapyd使用硒和webdriver進行抓取時,蜘蛛與命令「scrapy crawl myspider」當我部署使用scrapyd,最後用捲曲和scrapyd API安排它它觸發意想不到的關鍵字參數「_job」spider = cls(* args,** kwargs)TypeError:__init __()得到了一個意外的關鍵字參數'_job'
這裏是我的蜘蛛代碼
#!G:\python-2-7
import scrapy
from scrapy.spider import BaseSpider
from selenium import webdriver
from scrapy.http import TextResponse
import time
from time import sleep
import pickle
import math
from math import floor
from thevillages.items import ThevillagesItem
import MySQLdb
import sys
import json
class VillageSpider(BaseSpider):
name = 'village'
allowed_domains = ["example.com"]
start_urls = ['https://www.example.com/']
def __init__(self, *args, **kwargs):
super(VillageSpider, self).__init__(*args, **kwargs)
self.driver = webdriver.Firefox()
# def __init__(self):
def parse(self, response):
self.driver.get(response.url)
看看下面
一塊或錯誤日誌2017-10-17 17:58:05 [twisted] CRITICAL: Unhandled error in Deferred:
2017-10-17 17:58:05 [twisted] CRITICAL:
Traceback (most recent call last):
File "g:\python-2-7\lib\site-packages\twisted\internet\defer.py", line 1386, in _inlineCallbacks
result = g.send(result)
File "g:\python-2-7\lib\site-packages\scrapy\crawler.py", line 95, in crawl
six.reraise(*exc_info)
File "g:\python-2-7\lib\site-packages\scrapy\crawler.py", line 76, in crawl
self.spider = self._create_spider(*args, **kwargs)
File "g:\python-2-7\lib\site-packages\scrapy\crawler.py", line 99, in _create_spider
return self.spidercls.from_crawler(self, *args, **kwargs)
File "g:\python-2-7\lib\site-packages\scrapy\spiders\__init__.py", line 54, in from_crawler
spider = cls(*args, **kwargs)
TypeError: __init__() got an unexpected keyword argument '_job'
和scrapyd 1.2.0 –
@Tarun Lalwani感謝幫幫我。我得到了一個固定的,問題不是我在代碼粘貼問題,它也正常工作。在scrapyd部署的項目是緩存或我使用delproject.json scrapyd api刪除項目然後再部署它然後它開始工作正常。你也給出的答案是正在工作。謝謝 –