Comments (35)
This project is pretty good. If you can join Timed task, it will be a more complete project.
Come on!
from scrapydweb.
- You can select multiple crawlers when you want to set a timed task. You can select multiple crawlers to start at each time period after selecting the time point.
Imagine that I have 100 crawlers in a project. This feature does provide convenience and management.
2, as above, because there are a hundred crawlers, so can you have the function of custom label, that is, I can use the label screen and view the running status of the specified category of crawlers, and timed tasks
Thanks!
from scrapydweb.
Or modify the code below to vm.versions = ['default: the latest version'].concat(obj.versions);
from scrapydweb.
Or modify the code below to
vm.versions = ['default: the latest version'].concat(obj.versions);
scrapydweb/scrapydweb/templates/scrapydweb/schedule.htmlLine 826 in 560e998
vm.versions = obj.versions;
It works, thanks
from scrapydweb.
add a image of docker, please
@seozed
Check out the docker image created by @luzihang123.
from scrapydweb.
2、Online packaging and deployment
Have you tried the Projects > Deploy page?
from scrapydweb.
I know what you said, but this method should be manually packaged into egg, and then uploaded, and can not directly package the source project into egg, automatically upload it?
from scrapydweb.
I know what you said, but this method should be manually packaged into egg, and then uploaded, and can not directly package the source project into egg, automatically upload it?
Ok, I would try to figure out a better way to eggify local projects.
from scrapydweb.
You can refer to https://github.com/Gerapy/Gerapy
from scrapydweb.
v0.9.9: Add auto eggifying
from scrapydweb.
Wow, the response is fast! Expecting the same timing tasks
from scrapydweb.
You can select multiple crawlers when you want to set a timed task. You can select multiple crawlers to start at each time period after selecting the time point.
Imagine that I have 100 crawlers in a project. This feature does provide convenience and management.
You mean there are 100 spiders in a project and you want to schedule some of them to run periodically?
as above, because there are a hundred crawlers, so can you have the function of custom label, that is, I can use the label screen and view the running status of the specified category of crawlers, and timed tasks
What about labeling some related jobs with a specific jobid?
from scrapydweb.
You mean there are 100 spiders in a project and you want to schedule some of them to run periodically?
No, Simultaneously select multiple spiders to set one timed task.
What about labeling some related jobs with a specific jobid?
Labels will be better because they can be more visualized according to their own classification
from scrapydweb.
OK, I would take it into account when implementing this feature. Thanks for your advice!
from scrapydweb.
v1.0.0rc1: Add Email Notice, with multi-triggers provided, including:
-
ON_JOB_RUNNING_INTERVAL
-
ON_JOB_FINISHED
-
When reaching the threshold of specific kind of log: ['CRITICAL', 'ERROR', 'WARNING', 'REDIRECT', 'RETRY', 'IGNORE'], in the meanwhile, you can ask ScrapydWeb to stop/forcestop current job automatically.
Get it via the pip install scrapydweb==1.0.0rc1
command.
Email sample:
from scrapydweb.
ERROR in utils: !!!!! ConnectionError HTTPConnectionPool(host='127.0.0.1', port=6800): Max retries exceeded with url: /jobs (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x7f6111cae9b0>: Failed to establish a new connection: [Errno 111] Connection refused',))
from scrapydweb.
ERROR in utils: !!!!! ConnectionError HTTPConnectionPool(host='127.0.0.1', port=6800): Max retries exceeded with url: /jobs (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x7f6111cae9b0>: Failed to establish a new connection: [Errno 111] Connection refused',))
Please new an issue with details.
from scrapydweb.
timed task +1
from scrapydweb.
v1.2.0: Support Timer Tasks to schedule a spider run periodically
from scrapydweb.
While adding a timer task, I have to choose a version, but actually I only want to use the latest version so that if the project is updated, I don't need to update my task accordingly.
from scrapydweb.
OK, it would be fixed in the future release.
For the time being, go to the Jobs page and click either multinode or Start button to work around.
from scrapydweb.
The modification has been commited.
from scrapydweb.
For Logs categorization, can the same spiders distributed on different scrapyd be aggregated? @my8100
from scrapydweb.
@heave-Rother
For the time being, you can switch to a specific page of the neighboring node
with the help of Node Scroller and Node Skipping.
If that cannot satisfy your need, could you draw a picture to show me your idea?
from scrapydweb.
@my8100 ok,
from scrapydweb.
- What would you use the checkboxes in the drop-down list for?
Did you notice the checkboxes for nodes in the Servers page? - Stats aggregation would be followed up in PR #72.
from scrapydweb.
@my8100
yes, I want to choose the server I want to join the statistics.
from scrapydweb.
@heave-Rother
I see. Thanks for your suggestion.
from scrapydweb.
add a image of docker, please
from scrapydweb.
感谢作者,这是我找到的最好的爬虫集群操作平台。提几个需求:
1,给每个node加描述,方便自己看。
2,通过手机短信发送报警信息。
3,如何支持基于scrapy-redis的分布式爬虫的配置、启动?
from scrapydweb.
@devxiaosong
Replied in #107.
from scrapydweb.
Please add a short tutorial on how to switch from the flask development server to a production server with https enabled using letsencrypt. It would be much appreciated.
from scrapydweb.
Please add a short tutorial on how to switch from the flask development server to a production server with https enabled using letsencrypt. It would be much appreciated.
Please try it out and share your result.
Line 116 in 6b9663b
scrapydweb/scrapydweb/default_settings.py
Lines 82 to 92 in 6b9663b
from scrapydweb.
可以在页面上动态添加/删除 scrapyd 节点吗
from scrapydweb.
可以在页面上动态添加/删除 scrapyd 节点吗
Editing Scrapyd servers via GUI is not supported.
from scrapydweb.
Related Issues (20)
- project dependices package version incompatible HOT 3
- Not able to see stats section of the job HOT 1
- scrapydweb failed to run on python 3.8 HOT 5
- 启动报错:sqlite3.OperationalError: no such table: metadata HOT 13
- Is it possible to run multiple spider at the same time in a tmux machine with scrapydweb automatically
- items Oops! Something went wrong. HOT 1
- scrapydweb fresh install won't run HOT 8
- APScheduler 3.10 causing 500 errors HOT 2
- How to Change Timezone of scrapydweb? HOT 3
- Which scrapyd image you use? HOT 2
- Clean install on clean Ubuntu VM. Whatever I do it is not working. HOT 2
- Docker compose scrapdweb with scrapyd the log url use docker name
- Processes dont stop after finishing HOT 1
- v1.4.1 submit cron job can't run HOT 1
- ('Connection aborted.', timeout('timed out',))
- ERROR: Package 'scrapydweb' requires a different Python: HOT 4
- Error while installing scrapydweb HOT 2
- spiders are closed but showing as running/warning in the tasks page
- web界面可以使用中文吗? HOT 1
- DATABASE_URL配置连接域名:端口的mysql失败 HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from scrapydweb.