python爬虫scrapy之如何同时执行多个scrapy爬行任务

来源:https://www.shanhubei.com/archives/23601.html

1、顺序执行:

from scrapy.cmdline import execute
 
execute(['scrapy','crawl','httpbin'])

2、同时进行

    setting = get_project_settings()
    process = CrawlerProcess(setting)
    didntWorkSpider = ['sample']
    workSpider = ['gochinaz', 'gochinaz2', 'gochinaz3', 'gochinaz4', 'gochinaz5', 'gochinaz6', 'gochinaz7', 'gochinaz8']

    print("运行中...")
    for spider_name in process.spiders.list():
        if spider_name in workSpider:
            print("Running spider %s" % (spider_name))
            process.crawl(spider_name)
    process.start()

posted @   珊瑚贝博客  阅读(230)  评论(1编辑  收藏  举报
相关博文:
阅读排行:
· 全程不用写代码,我用AI程序员写了一个飞机大战
· DeepSeek 开源周回顾「GitHub 热点速览」
· 记一次.NET内存居高不下排查解决与启示
· MongoDB 8.0这个新功能碉堡了,比商业数据库还牛
· .NET10 - 预览版1新功能体验(一)
点击右上角即可分享
微信分享提示