在scrapy_splash中加載本地Cookies


Splash→Scrapy:

首先,lua腳本中要返回Cookies:

return {cookies = splash:get_cookies()}

然后,在spider中調用response.cookiejar即可得到返回的Cookies

Scrapy→Splash

把Cookies作為參數傳遞到SplashRequest()中:

def start_requests(self):
        local = "D:\\...\\VertifiedCookies.txt"
        with open(local, 'r+') as f:
            cookies = eval(f.read())
        url = 'https://www.example.com'
        script = """
        function main(splash)
            splash:clear_cookies()
            splash:init_cookies(splash.args.cookies)
            assert(splash:go(splash.args.url))
            assert(splash:wait(splash.args.wait))
            return {cookies = splash:get_cookies(),html = splash:html()}
        end
        """
        yield SplashRequest(url, self.parse, endpoint='execute', args={'lua_source':script,'wait':1}, cookies=cookies)

傳入數據的參數類型與scrapy.Request中的cookies相同:


免責聲明!

本站轉載的文章為個人學習借鑒使用,本站對版權不負任何法律責任。如果侵犯了您的隱私權益,請聯系本站郵箱yoyou2525@163.com刪除。



 
粵ICP備18138465號   © 2018-2025 CODEPRJ.COM