Scrapy爬虫实战三:获取代理

来源:互联网 发布:座天使兽网络侦探 编辑:程序博客网 时间:2024/06/05 10:42

本文项目采用python3.6版本语言,利用scrapy框架进行爬取。

该项目实现的功能是获取http://www.proxy360.cn和http://www.xicidaili.com网站中的代理信息,由于网站设有反爬虫机制,网站是通过浏览器发送过来的User-Agent的值来确认浏览器身份的,所以为了避免被查出是爬虫,所以该项目中修改了USER_AGENT的值,关于常见反爬虫机制请参照本博客“Scrapy爬虫实战五:爬虫攻防”的内容。由于爬取http://www.xicidaili.com的内容较多,在验证代理是否可用的过程中使用了多线程。

下面是本次项目的目录结构:

----getProxy

----getProxy

----spiders

__init__.py

Proxy360Spider.py

XiciDailiSpider.py

__init__.py

items.py

pipelines.py

settings.py

userAgents.py

scrapy.cfg

testProxy.py





上述目录结构中,没有后缀名的为文件夹,有后缀的为文件。我们需要修改只有Proxy360Spider.py(XiciDailiSpider.py、items.py、pipelines.py、settings.py这四个文件。

其中items.py决定爬取哪些项目,Proxy360Spider.py(XiciDailiSpider.py决定怎么爬,setting.py决定由谁去处理爬取的内容,pipelines决定爬取后

内容怎样处理。由于我们爬取的两个代理网站的源代码不类似,所以我们写了Proxy360Spider.py(XiciDailiSpider.py)两个文件,目录中的两个

__inti__.py文件都是空文件,保留这两个文件主要是为了让他们所在的文件夹可以作为Python的模块使用。目录中的scrapy.cfg为配置文件,testProxy.py是测试爬取出来的

代理是否可用的。


1、选择爬取的项目items.py

#决定爬取哪些项目import scrapyclass GetProxyItem(scrapy.Item):    ip=scrapy.Field()    port=scrapy.Field()    type=scrapy.Field()    location=scrapy.Field()    protocol=scrapy.Field()    source=scrapy.Field()    id=scrapy.Field()#用来保存记录条数

2.1、定义怎样爬取,爬取360代理的文件Proxy360Spider.py

#定义如何爬取import scrapyfrom getProxy.items import GetProxyItemclass Proxy360Spider(scrapy.Spider):    name="proxy360Spider"    allowed_domains=['proxy360.cn']    nations=['Brazil','China','America','Taiwan','Japan','Thailand','Vietnam','bahrein']    start_urls=[]    count=0    for nation in nations:        start_urls.append('http://www.proxy360.cn/Region/'+nation)    def parse(self,response):                subSelector=response.xpath('//div[@class="proxylistitem" and @name="list_proxy_ip"]')        items=[]                for sub in subSelector:            item=GetProxyItem()            item['ip']=sub.xpath('.//span[1]/text()').extract()[0]            item['port']=sub.xpath('.//span[2]/text()').extract()[0]            item['type']=sub.xpath('.//span[3]/text()').extract()[0]            item['location']=sub.xpath('.//span[4]/text()').extract()[0]            item['protocol']='HTTP'            item['source']='proxy360'            self.count+=1            item['id']=str(self.count)            items.append(item)        return items

由于360代理中包含多个地区的代理,首先找寻多个地区代理url的规律,可以这些地区的url都是由“http://www.proxy360.cn/Region/”加上地区名构成的,

任意点开一个url,打开网页源代码,如下图所示:



可以发现我们需要的代理信息都在<divclass="proxylistitem"name="list_proxy_ip">这个标签下,用xpath选择器先得到所有的这种标签,然后再一层一层的得到我们需要的信息。


2.1、定义怎样爬取,爬取西刺免费代理IP的文件XiciDailiSpider.py

import scrapyfrom getProxy.items import GetProxyItemclass XiciDailiSpider(scrapy.Spider):    name='xiciDailiSpider'    allowed_domains=['xicidaili.com']    wds=['nn','nt','wn','wt']    pages=20    start_urls=[]    count=0    for type in wds:        for i in range(1,pages+1):            start_urls.append('http://www.xicidaili.com/'+type+'/'+str(i))    def parse(self,response):        subSelector=response.xpath('//tr[@class=""]|//tr[@class="odd"]')        items=[]        for sub in subSelector:            item=GetProxyItem()            item['ip']=sub.xpath('.//td[2]/text()').extract()[0]            item['port']=sub.xpath('.//td[3]/text()').extract()[0]            item['type']=sub.xpath('.//td[5]/text()').extract()[0]            if sub.xpath('.//td[4]/a/text()'):                item['location']=sub.xpath('//td[4]/a/text()').extract()[0]            else:                item['location']=sub.xpath('.//td[4]/text()').extract()[0]            item['protocol']=sub.xpath('.//td[6]/text()').extract()[0]            item['source']='xicidaili'            self.count+=1            item['id']=str(self.count)            items.append(item)        return items            

同样爬取西刺代理的url也是有多个,第一步仍然是找寻url规律,读者去http://www.xicidaili.com/网站中多点几下,关注url的变化就能得出规律,这里不再赘述。


查看网页源代码可以发现,我们需要的代理信息都在<trclass="">或者<trclass="odd">这两个标签下,读者重点看代码中是如何一层一层找到我们需要的信息的。

3、保存爬取的结果pipelines.py

#保存爬取结果import timeclass GetProxyPipeline(object):    def process_item(self,item,spider):        today= time.strftime('%Y-%m-%d',time.localtime())        fileName='proxy'+today+'.txt'        with open(fileName,'a') as fp:            fp.write('记录'+item['id'].strip()+':'+'\t')            fp.write(item['ip'].strip()+'\t')            fp.write(item['port'].strip()+'\t')            fp.write(item['protocol'].strip()+'\t')            fp.write(item['type'].strip()+'\t')            fp.write(item['location'].strip()+'\t')            fp.write(item['source'].strip()+'\t\n')                    return item

结果保存为.txt的文件,文件以proxy加当天日期命名。


4、分派任务的settings.py

from getProxy import userAgentsBOT_NAME='getProxy'SPIDER_MODULES=['getProxy.spiders']NEWSPIDER_MODULE='getProxy.spiders'USER_AGENT=userAgents.pcUserAgent.get('Firefox 4.0.1 – Windows')ITEM_PIPELINES={'getProxy.pipelines.GetProxyPipeline':300}


这里修改了USER_AGENT,导入userAgents模块,下面给出userAgents模块代码:

pcUserAgent = {"safari 5.1 – MAC":"User-Agent:Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50","safari 5.1 – Windows":"User-Agent:Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50","IE 9.0":"User-Agent:Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Trident/5.0);","IE 8.0":"User-Agent:Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0)","IE 7.0":"User-Agent:Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)","IE 6.0":"User-Agent: Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1)","Firefox 4.0.1 – MAC":"User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:2.0.1) Gecko/20100101 Firefox/4.0.1","Firefox 4.0.1 – Windows":"User-Agent:Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1","Opera 11.11 – MAC":"User-Agent:Opera/9.80 (Macintosh; Intel Mac OS X 10.6.8; U; en) Presto/2.8.131 Version/11.11","Opera 11.11 – Windows":"User-Agent:Opera/9.80 (Windows NT 6.1; U; en) Presto/2.8.131 Version/11.11","Chrome 17.0 – MAC":"User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_0) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 Safari/535.11","Maxthon":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Maxthon 2.0)","Tencent TT":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; TencentTraveler 4.0)","The World 2.x":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)","The World 3.x":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; The World)","sogou 1.x":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Trident/4.0; SE 2.X MetaSr 1.0; SE 2.X MetaSr 1.0; .NET CLR 2.0.50727; SE 2.X MetaSr 1.0)","360":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; 360SE)","Avant":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Avant Browser)","Green Browser":"User-Agent: Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)"}mobileUserAgent = {"iOS 4.33 – iPhone":"User-Agent:Mozilla/5.0 (iPhone; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5","iOS 4.33 – iPod Touch":"User-Agent:Mozilla/5.0 (iPod; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5","iOS 4.33 – iPad":"User-Agent:Mozilla/5.0 (iPad; U; CPU OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) Version/5.0.2 Mobile/8J2 Safari/6533.18.5","Android N1":"User-Agent: Mozilla/5.0 (Linux; U; Android 2.3.7; en-us; Nexus One Build/FRF91) AppleWebKit/533.1 (KHTML, like Gecko) Version/4.0 Mobile Safari/533.1","Android QQ":"User-Agent: MQQBrowser/26 Mozilla/5.0 (Linux; U; Android 2.3.7; zh-cn; MB200 Build/GRJ22; CyanogenMod-7) AppleWebKit/533.1 (KHTML, like Gecko) Version/4.0 Mobile Safari/533.1","Android Opera ":"User-Agent: Opera/9.80 (Android 2.3.4; Linux; Opera Mobi/build-1107180945; U; en-GB) Presto/2.8.149 Version/11.10","Android Pad Moto Xoom":"User-Agent: Mozilla/5.0 (Linux; U; Android 3.0; en-us; Xoom Build/HRI39) AppleWebKit/534.13 (KHTML, like Gecko) Version/4.0 Safari/534.13","BlackBerry":"User-Agent: Mozilla/5.0 (BlackBerry; U; BlackBerry 9800; en) AppleWebKit/534.1+ (KHTML, like Gecko) Version/6.0.0.337 Mobile Safari/534.1+","WebOS HP Touchpad":"User-Agent: Mozilla/5.0 (hp-tablet; Linux; hpwOS/3.0.0; U; en-US) AppleWebKit/534.6 (KHTML, like Gecko) wOSBrowser/233.70 Safari/534.6 TouchPad/1.0","Nokia N97":"User-Agent: Mozilla/5.0 (SymbianOS/9.4; Series60/5.0 NokiaN97-1/20.0.019; Profile/MIDP-2.1 Configuration/CLDC-1.1) AppleWebKit/525 (KHTML, like Gecko) BrowserNG/7.1.18124","Windows Phone Mango":"User-Agent: Mozilla/5.0 (compatible; MSIE 9.0; Windows Phone OS 7.5; Trident/5.0; IEMobile/9.0; HTC; Titan)","UC":"User-Agent: UCWEB7.0.2.37/28/999","UC standard":"User-Agent: NOKIA5700/ UCWEB7.0.2.37/28/999","UCOpenwave":"User-Agent: Openwave/ UCWEB7.0.2.37/28/999","UC Opera":"User-Agent: Mozilla/4.0 (compatible; MSIE 6.0; ) Opera/UCWEB7.0.2.37/28/999"}

5、配置文件scrapy.cfg

[settings]default=getProxy.settings[deploy]project=getProxy


6、怎么运行

cmd->cd 将文件调到我们项目所在的这一层文件,也就是上面目录结构中scrapy.cfg所在的这一层文件夹,然后输入命令:scrapy crawl proxy360Spider

或者scrapy crawl xiciDailiSpider

这里的proxy360Spider是我们Proxy360SpiderSpider类中name="proxy360Spider"的值,更改name的值输入的命令也将更改,xiciDailiSpider也是同理。

我们写了两个网站的爬虫,所以要用依次输入上面的两个命令,主要是因为360代理中的代理ip太少,所以又多爬取了一个网站。

7、测试爬取出的代理是否可用testProxy.py

from urllib import requestimport reimport threadingimport timeclass TestProxy(object):    def __init__(self):        self.sFile=r'proxy'+time.strftime('%Y-%m-%d',time.localtime())+'.txt'   #     self.sFile=r'alive.txt'        self.dFile=r'alive_.txt'        self.URL=r'http://www.baidu.com/'        self.threads=10        self.timeout=3        self.regex=re.compile(r'baidu.com')        self.aliveList=[]        self.run()    def run(self):        with open(self.sFile,'r') as fp:            lines=fp.readlines()            line=lines.pop()            while lines:                for i in range(self.threads):                    try:                        t=threading.Thread(target=self.linkWithProxy,args=(line,))                        t.start()                        if lines:                            line=lines.pop()                        else:                            continue                    except:                        print('创建线程失败!\n')                                with open(self.dFile,'a') as fp:            for i in range(len(self.aliveList)):                fp.write(self.aliveList[i])    def linkWithProxy(self,line):        lineList=line.split('\t')        protocol=lineList[3].lower()        server=protocol+r'://'+lineList[1]+':'+lineList[2]        opener=request.build_opener(request.ProxyHandler({protocol:server}))        request.install_opener(opener)        try:            response=request.urlopen(self.URL,timeout=self.timeout)        except:            print('%s connect failed!\n' %server)            return        else:            try:                strRe=response.read()            except:                print('%s connect failed!\n' %server)                return            if self.regex.search(str(strRe)):                print('%s connect success!!!!!!!\n' %server)                self.aliveList.append(line)if __name__=='__main__':    tp=TestProxy()            

这里是读取proxy+当天日期的.txt文件,也就是上面爬虫跑出来的结果,然后将可用的ip写在alive_.txt文件中,这些运行的结果都是保存在项目的根目录中。



本博客有参考《Python网络爬虫实战》一书,该书采用的是python2.x在Linux系统下运行的,采用python3.x在windows下运行的可以参考本博客。