# spiderkeeper **Repository Path**: zegege/spiderkeeper ## Basic Information - **Project Name**: spiderkeeper - **Description**: No description available - **Primary Language**: Unknown - **License**: Not specified - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2021-01-18 - **Last Updated**: 2024-12-27 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # spiderkeeper 使用开源的spiderkeeper,然后做了自己的修改 把之前spiderkeeper添加任务简化并增加了重试机制 现在添加任务非常稳定(需要启动时候是 --no--auth )。 # use git clone ssh://git@github.com:wem603947175/spiderkeeper.git cd SpiderKeeper pip3 install . ----------------------------------------------- ./spider_keeper.sh 接着上传自己的egg包 ----------------------------------------------- 开启多任务是调用api接口: import requests spiderkeeper_url = 'http://127.0.0.1:22250/api/projects/2/jobs' data = { 'spider_name': "kuasheng_spider_v2", 'spider_arguments': "-----", 'run_type': 'onetime', 'priority': 0 } print(data) res = requests.post(spiderkeeper_url, data=data) print(res.content) if res.text == 'true\n': print('ok') else: print('fail') ----------------------------------------------- # SpiderKeeper [![Latest Version](http://img.shields.io/pypi/v/SpiderKeeper.svg)](https://pypi.python.org/pypi/SpiderKeeper) [![Python Versions](http://img.shields.io/pypi/pyversions/SpiderKeeper.svg)](https://pypi.python.org/pypi/SpiderKeeper) [![The MIT License](http://img.shields.io/badge/license-MIT-blue.svg)](https://github.com/DormyMo/SpiderKeeper/blob/master/LICENSE) A scalable admin ui for spider service ## Features - Manage your spiders from a dashboard. Schedule them to run automatically - With a single click deploy the scrapy project - Show spider running stats - Provide api Current Support spider service - [Scrapy](https://github.com/scrapy/scrapy) ( with [scrapyd](https://github.com/scrapy/scrapyd)) ## Screenshot ![job dashboard](https://raw.githubusercontent.com/DormyMo/SpiderKeeper/master/screenshot/screenshot_1.png) ![periodic job](https://raw.githubusercontent.com/DormyMo/SpiderKeeper/master/screenshot/screenshot_2.png) ![running stats](https://raw.githubusercontent.com/DormyMo/SpiderKeeper/master/screenshot/screenshot_3.png) ## Getting Started ### Installing ``` pip install spiderkeeper ``` ### Deployment ``` spiderkeeper [options] Options: -h, --help show this help message and exit --host=HOST host, default:0.0.0.0 --port=PORT port, default:5000 --username=USERNAME basic auth username ,default: admin --password=PASSWORD basic auth password ,default: admin --type=SERVER_TYPE access spider server type, default: scrapyd --server=SERVERS servers, default: ['http://localhost:6800'] --database-url=DATABASE_URL SpiderKeeper metadata database default: sqlite:////home/souche/SpiderKeeper.db --no-auth disable basic auth -v, --verbose log level example: spiderkeeper --server=http://localhost:6800 ``` ## Usage ``` Visit: - web ui : http://localhost:5000 1. Create Project 2. Use [scrapyd-client](https://github.com/scrapy/scrapyd-client) to generate egg file scrapyd-deploy --build-egg output.egg 2. upload egg file (make sure you started scrapyd server) 3. Done & Enjoy it - api swagger: http://localhost:5000/api.html ``` ## TODO - [ ] Job dashboard support filter - [x] User Authentication - [ ] Collect & Show scrapy crawl stats - [ ] Optimize load balancing ## Versioning We use [SemVer](http://semver.org/) for versioning. For the versions available, see the [tags on this repository](https://github.com/DormyMo/SpiderKeeper/tags). ## Authors - *Initial work* - [DormyMo](https://github.com/DormyMo) See also the list of [contributors](https://github.com/DormyMo/SpiderKeeper/contributors) who participated in this project. ## License This project is licensed under the MIT License - see the [LICENSE.md](LICENSE.md) file for details ## Contributing Contributions are welcomed!