2012-04-24 106 views
3

我已经设置了我的PATH变量,我想我正在配置一切正确。但是,当我在我的startproject命令文件夹中运行“scrapy爬行DMOZ”,我得到这些错误消息:scrapy新手:教程。运行scrapy爬行时出错dmoz

c:\matt\testing\dmoz>scrapy crawl dmoz 
2012-04-24 18:12:56-0400 [scrapy] INFO: Scrapy 0.14.0.2841 started (bot: dmoz) 
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole,   
CloseSpider, WebService, CoreStats, SpiderState 
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled downloader middlewares:  
HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware, 
faultHeadersMiddleware, RedirectMiddleware, CookiesMiddleware, 
HttpCompressionMiddleware, ChunkedTransferMiddleware, DownloaderStats 
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled spider middlewares: 
HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddware 
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled item pipelines: 
Traceback (most recent call last): 
File "c:\Python27\Scripts\scrapy", line 4, in <module> 
execute() 
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7- 
win32.egg\scrapy\cmdline.py", line 132, in execute 
_run_print_help(parser, _run_command, cmd, args, opts) 
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7- 
win32.egg\scrapy\cmdline.py", line 97, in _run_print_help 
func(*a, **kw) 
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7- 
win32.egg\scrapy\cmdline.py", line 139, in _run_command 
cmd.run(args, opts) 
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7- 
win32.egg\scrapy\commands\crawl.py", line 43, in run 
spider = self.crawler.spiders.create(spname, **opts.spargs) 
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7- 
win32.egg\scrapy\spidermanager.py", line 43, in create 
raise KeyError("Spider not found: %s" % spider_name) 
KeyError: 'Spider not found: dmoz' 

有谁知道什么可能是怎么回事?

+2

显示你的目录结构,你的蜘蛛类的源码和设置模块 – warvariuc 2012-04-25 07:27:08

+0

请选择一个最好的答案! – realtebo 2013-06-08 10:17:32

+0

此外,本教程未能提及您还需要此库: https://github.com/scrapy/queuelib – 2013-11-19 02:25:56

回答

0

尝试在命令行

C:\Users\Akhtar Wahab> python # any any directory path 

,如果它工作

尝试

scrapy version 

如果还工作

然后确保你

做出scrapy项目
scrapy startproject name 

如果上述一切工作正适合你

然后确保你是在一个目录下scrapy.cfg存在刮板运行命令。

12

我也有这个问题。

这是因为scrapy教程会要求您将您创建的蜘蛛放置在/dmoz/spiders/中,但scrapy正在寻找tutorial/tutorial/spiders

dmoz_spider.py保存在tutorial/tutorial/spiders中,爬网应该可以工作。

+0

谢谢!!!!!!为什么作者没有选择最佳答案? – realtebo 2013-06-08 10:16:59