參照官方文檔:https://doc.scrapy.org/en/latest/topics/practices.html,認真學習文檔才是正解
1.安裝pyinstaller
2.安裝pywin32
3.安裝其他模塊
4.在爬蟲項目里進行相關操作,參照博友:https://blog.csdn.net/la_vie_est_belle/article/details/79017358
4.1在scrapy.cfg文件同路徑下創建s_spider.py
4.2寫入相關代碼
# -*- coding: utf-8 -*- from scrapy.crawler import CrawlerProcess from scrapy.utils.project import get_project_settings # 這里是必須引入的 import robotparser import scrapy.spiderloader import scrapy.statscollectors import scrapy.logformatter import scrapy.dupefilters import scrapy.squeues import scrapy.extensions.spiderstate import scrapy.extensions.corestats import scrapy.extensions.telnet import scrapy.extensions.logstats import scrapy.extensions.memusage import scrapy.extensions.memdebug import scrapy.extensions.feedexport import scrapy.extensions.closespider import scrapy.extensions.debug import scrapy.extensions.httpcache import scrapy.extensions.statsmailer import scrapy.extensions.throttle import scrapy.core.scheduler import scrapy.core.engine import scrapy.core.scraper import scrapy.core.spidermw import scrapy.core.downloader import scrapy.downloadermiddlewares.stats import scrapy.downloadermiddlewares.httpcache import scrapy.downloadermiddlewares.cookies import scrapy.downloadermiddlewares.useragent import scrapy.downloadermiddlewares.httpproxy import scrapy.downloadermiddlewares.ajaxcrawl import scrapy.downloadermiddlewares.chunked import scrapy.downloadermiddlewares.decompression import scrapy.downloadermiddlewares.defaultheaders import scrapy.downloadermiddlewares.downloadtimeout import scrapy.downloadermiddlewares.httpauth import scrapy.downloadermiddlewares.httpcompression import scrapy.downloadermiddlewares.redirect import scrapy.downloadermiddlewares.retry import scrapy.downloadermiddlewares.robotstxt import scrapy.spidermiddlewares.depth import scrapy.spidermiddlewares.httperror import scrapy.spidermiddlewares.offsite import scrapy.spidermiddlewares.referer import scrapy.spidermiddlewares.urllength import scrapy.pipelines import scrapy.core.downloader.handlers.http import scrapy.core.downloader.contextfactory # 自己項目用到的 #import scrapy.pipelines.images # 用到圖片管道 import openpyxl # 用到openpyxl庫 process = CrawlerProcess(get_project_settings()) # 'sk' is the name of one of the spiders of the project. process.crawl('sk') process.start() # the script will block here until the crawling is finished
4.3在s_spider.py目錄下:shift+右鍵,然后點擊‘’在此處打開命令窗口‘’,輸入:pyinstaller crawl.py,生成dist,build(可刪)和crawl.spec(可刪)。
4.4在s_spider.exe目錄下創建文件夾scrapy,然后到自己安裝的scrapy文件夾中把VERSION和mime.types兩個文件復制到剛才創建的scrapy文件夾中。
4.5重新打包運行.exe即可,需要爬蟲支持,不能單獨運行。