selenium登录知乎并且输出一些首页的标题demo

来源:互联网 发布:android c语言开发 编辑:程序博客网 时间:2024/06/05 02:03

转载自http://blog.csdn.net/uselym/article/details/52525025
主要用到了selenium进行登录验证,用cookies来给scrapy,然后用自带的xpath解析网页,其实似乎应该selenium登录解析放在中间件里的,不过这里先完成功能而已吧

# -*- coding: utf-8 -*-import scrapyfrom selenium import webdriverclass ZhihuSpider(scrapy.Spider):    name = "zhihu"    allowed_domains = ["zhihu.com"]    start_urls = (        'https://www.zhihu.com/',    )    def get_cookies(self):        driver = webdriver.Chrome()        driver.get(self.start_urls[0])        driver.find_element_by_link_text(u"登录").click()        driver.find_element_by_name("account").clear()        driver.find_element_by_name("account").send_keys("your username") #修改为自己的用户名        driver.find_element_by_name("password").clear()        driver.find_element_by_name("password").send_keys("keys") #修改为自己的密码        SignInURL = u"https://www.zhihu.com/#signin"        try:            if driver.find_element_by_id('captcha'):                while True:                    if not SignInURL == driver.current_url:                        break                    pass                pass        finally:            if SignInURL == driver.current_url:                driver.find_element_by_css_selector("button.sign-button.submit").click()            cookies = driver.get_cookies()            driver.close()            print cookies            return cookies    def after_login(self, response):        sel = scrapy.Selector(response)        #print response.body        for i in range(1,10):            xml = r'//*[@id="feed-%d"]/div[1]/div[2]/div[2]/h2/a/text()' %(i)            titile = sel.xpath(xml).extract()            if len(titile):                print str(titile[0])    def parse(self, response):        return scrapy.Request(url=self.start_urls[0], cookies=self.get_cookies(), callback=self.after_login)
原创粉丝点击