@@ -60,7 +60,7 @@ def __init__(self, settings, scrapyrt_manager):
6060
6161 def crawl (self , spidercls , * args , ** kwargs ):
6262 if isinstance (spidercls , six .string_types ):
63- spidercls = self .spiders .load (spidercls )
63+ spidercls = self .spider_loader .load (spidercls )
6464 # creating our own crawler that will allow us to disable start requests easily
6565 crawler = ScrapyrtCrawler (
6666 spidercls , self .settings , self .scrapyrt_manager .start_requests )
@@ -159,7 +159,7 @@ def crawl(self, *args, **kwargs):
159159 dfd = self .crawler_process .crawl (self .spider_name , * args , ** kwargs )
160160 except KeyError as e :
161161 # Spider not found.
162- raise Error ('404' , message = e . message )
162+ raise Error ('404' , message = str ( e ) )
163163 dfd .addCallback (self .return_items )
164164 return dfd
165165
@@ -245,7 +245,7 @@ def collect_dropped(self, item, response, exception, spider):
245245 if spider is self .crawler .spider :
246246 self .items_dropped .append ({
247247 "item" : item ,
248- "exception" : exception . message ,
248+ "exception" : str ( exception ) ,
249249 "response" : response
250250 })
251251
@@ -267,7 +267,8 @@ def create_spider_request(self, kwargs):
267267 try :
268268 req = Request (url , ** kwargs )
269269 except (TypeError , ValueError ) as e :
270- message = "Error while creating Scrapy Request, {}" .format (e .message )
270+ msg = "Error while creating Scrapy Request, {}"
271+ message = msg .format (str (e ))
271272 raise Error ('400' , message = message )
272273
273274 req .dont_filter = True
0 commit comments