按照上一篇文章中咱們將代碼放到遠程主機是經過拷貝或者git的方式,可是若是考慮到咱們又多臺遠程主機的狀況,這種方式就比較麻煩,那有沒有好用的方法呢?這裏其實能夠經過scrapyd,下面是這個scrapyd的github地址:https://github.com/scrapy/scrapydnode
當在遠程主機上安裝了scrapyd並啓動以後,就會再遠程主機上啓動一個web服務,默認是6800端口,這樣咱們就能夠經過http請求的方式,經過接口的方式管理咱們scrapy項目,這樣就不須要在一個一個電腦鏈接拷貝過着經過git,關於scrapyd官方文檔地址:http://scrapyd.readthedocs.io/en/stable/python
安裝scrapyd:pip install scrapydlinux
這裏我在另一臺ubuntu linux虛擬機中一樣安裝scrapy以及scrapyd等包,保證所要運行的爬蟲須要的包都完成安裝,這樣咱們就有了兩臺linux,包括上篇文章中咱們已經有的linux環境git
在這裏有個小問題須要注意,默認scrapyd啓動是經過scrapyd就能夠直接啓動,這裏bind綁定的ip地址是127.0.0.1端口是:6800,這裏爲了其餘虛擬機訪問講ip地址設置爲0.0.0.0
scrapyd的配置文件:/usr/local/lib/python3.5/dist-packages/scrapyd/default_scrapyd.confgithub
這樣咱們就能夠經過瀏覽器訪問:web
如何經過scrapyd部署項目,這裏官方文檔提供一個地址:https://github.com/scrapy/scrapyd-client,即經過scrapyd-client進行操做redis
這裏的scrapyd-client主要實現如下內容:json
咱們將咱們本地的scrapy項目中scrapy.cfg配置文件進行配置:ubuntu
咱們其實還能夠設置用戶名和密碼,不過這裏沒什麼必要,只設置了url
這裏設置url必定要注意:url = http://192.168.1.9:6800/addversion.json
最後的addversion.json不能少api
咱們在本地安裝pip install scrapy_client,安裝完成後執行:scrapyd-deploy
zhaofandeMBP:zhihu_user zhaofan$ scrapyd-deploy Packing version 1502177138 Deploying to project "zhihu_user" in http://192.168.1.9:6800/addversion.json Server response (200): {"node_name": "fan-VirtualBox", "status": "ok", "version": "1502177138", "spiders": 1, "project": "zhihu_user"} zhaofandeMBP:zhihu_user zhaofan$
看到status:200表示已經成功
listprojects.json列出上傳的項目列表
zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/listprojects.json {"node_name": "fan-VirtualBox", "status": "ok", "projects": ["zhihu_user"]} zhaofandeMBP:zhihu_user zhaofan$
listversions.json列出有某個上傳項目的版本
zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/listversions.json\?project\=zhihu_user {"node_name": "fan-VirtualBox", "status": "ok", "versions": ["1502177138"]} zhaofandeMBP:zhihu_user zhaofan$
schedule.json遠程任務的啓動
下面咱們啓動的三次就表示咱們啓動了三個任務,也就是三個調度任務來運行zhihu這個爬蟲
zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/schedule.json -d project=zhihu_user -d spider=zhihu {"node_name": "fan-VirtualBox", "status": "ok", "jobid": "97f1b5027c0e11e7b07a080027bbde73"} zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/schedule.json -d project=zhihu_user -d spider=zhihu {"node_name": "fan-VirtualBox", "status": "ok", "jobid": "99595aa87c0e11e7b07a080027bbde73"} zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/schedule.json -d project=zhihu_user -d spider=zhihu {"node_name": "fan-VirtualBox", "status": "ok", "jobid": "9abb1ba27c0e11e7b07a080027bbde73"} zhaofandeMBP:zhihu_user zhaofan$
同時當啓動完成後,咱們能夠經過頁面查看jobs,這裏由於我遠端服務器並無安裝scrapy_redis,因此顯示任務是完成了,我點開日誌並能看到詳細的日誌狀況:
這裏出錯的緣由就是我上面忘記在ubuntu虛擬機安裝scrapy_redis以及pymongo模塊,進行
pip install scrapy_redis pymongo安裝後從新啓動,就能夠看到已經在運行的任務,同時點開Log日誌也能看到爬取到的內容:
listjobs.json列出全部的jobs任務
上面是經過頁面顯示全部的任務,這裏是經過命令獲取結果
zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/listjobs.json\?project\=zhihu_user {"node_name": "fan-VirtualBox", "status": "ok", "running": [], "pending": [], "finished": [{"start_time": "2017-08-08 15:53:00.510050", "spider": "zhihu", "id": "97f1b5027c0e11e7b07a080027bbde73", "end_time": "2017-08-08 15:53:01.416139"}, {"start_time": "2017-08-08 15:53:05.509337", "spider": "zhihu", "id": "99595aa87c0e11e7b07a080027bbde73", "end_time": "2017-08-08 15:53:06.627125"}, {"start_time": "2017-08-08 15:53:10.509978", "spider": "zhihu", "id": "9abb1ba27c0e11e7b07a080027bbde73", "end_time": "2017-08-08 15:53:11.542001"}]} zhaofandeMBP:zhihu_user zhaofan$
cancel.json取消全部運行的任務
這裏能夠將上面啓動的全部jobs均可以取消:
zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/cancel.json -d project=zhihu_user -d job=0f5cdabc7c1011e7b07a080027bbde73 {"node_name": "fan-VirtualBox", "status": "ok", "prevstate": "running"} zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/cancel.json -d project=zhihu_user -d job=63f8e12e7c1011e7b07a080027bbde73 {"node_name": "fan-VirtualBox", "status": "ok", "prevstate": "running"} zhaofandeMBP:zhihu_user zhaofan$ curl http://192.168.1.9:6800/cancel.json -d project=zhihu_user -d job=63f8e12f7c1011e7b07a080027bbde73 {"node_name": "fan-VirtualBox", "status": "ok", "prevstate": "running"}
這樣當咱們再次經過頁面查看,就能夠看到全部的任務都是finshed狀態:
我相信看了上面這幾個方法你必定會以爲真不方便還須要輸入那麼長,因此有人替你幹了件好事把這些API進行的再次封裝:https://github.com/djm/python-scrapyd-api
該模塊可讓咱們直接在python代碼中進行上述那些api的操做
首先先安裝該模塊:pip install python-scrapyd-api
使用方法以下,這裏只演示了簡單的例子,其餘方法其實使用很簡單按照規則寫就行:
from scrapyd_api import ScrapydAPI scrapyd = ScrapydAPI('http://192.168.1.9:6800') res = scrapyd.list_projects() res2 = scrapyd.list_jobs('zhihu_user') print(res) print(res2)
Cancel a scheduled job
scrapyd.cancel('project_name', '14a6599ef67111e38a0e080027880ca6')
Delete a project and all sibling versions
scrapyd.delete_project('project_name')
Delete a version of a project
scrapyd.delete_version('project_name', 'version_name')
Request status of a job
scrapyd.job_status('project_name', '14a6599ef67111e38a0e080027880ca6')
List all jobs registered
scrapyd.list_jobs('project_name')
List all projects registered
scrapyd.list_projects()
List all spiders available to a given project
scrapyd.list_spiders('project_name')
List all versions registered to a given project
scrapyd.list_versions('project_name')
Schedule a job to run with a specific spider
scrapyd.schedule('project_name', 'spider_name')
Schedule a job to run while passing override settings
settings = {'DOWNLOAD_DELAY': 2}
Schedule a job to run while passing extra attributes to spider initialisationscrapyd.schedule('project_name', 'spider_name', extra_attribute='value')