爬虫京东网站

import requests
from requests.exceptions import RequestException
import re
import csv
import threadpool#需要先安装:pip install threadpool
#文件保存路径
path=‘E:\data.csv’
#访问网址
baseUrl = ‘https://maoyan.com/board/4?offset=

def parseOnePage(url):
try:
response = requests.get(url)
except RequestException:
print(url+‘url请求发生异常’)
return None
if response.status_code!=200:
print(‘状态码不为200,’+url+’ 响应异常’)
return None
else:#正常响应
html = response.text
pattern = re.compile(#正则表达式

. ?board-index.?>(\d+). ?class=“name”>?<a. ?>(.?). ?class=“star”>(.?)

. ?“releasetime”>(.?). ?class=“score”.?“integer”>(. ?).?“fraction”>(. ?).?’,
re.S)
res = re.findall(pattern,html)
# print(res)
for movie in res:
yield{
‘index’:movie[0].strip(),
‘name’:movie[1].strip(),
‘star’:movie[2].strip()[3:],
‘time’:movie[3].strip()[5:15],
‘score’:movie[4].strip() + movie[5].strip()
}
def writeToCSVFile(content):#将结果存为text文件
with open(path,‘a’,newline=’’) as f:
csv_writer = csv.writer(f,dialect=‘excel’)
csv_writer.writerow(content)
f.close();
def spider(offset):#爬取单个网页数据
url = baseUrl + str(offset)
for item in parseOnePage(url):
print(item.values())
writeToCSVFile(item.values())

def init():#写文件头
with open(path,‘w’,newline=’’) as f:
head = [‘排名’,‘电影名字’,‘主演’,‘上映时间’,‘评分’]
csv_writer = csv.writer(f, dialect=‘excel’)
csv_writer.writerow(head)
f.close()

def singelThread():#单线程抓取数据
init()
for offset in range(0,10):
spider(10*offset)

def multiThread():#多线程抓取数据
init()
pool = threadpool.ThreadPool(10)
requests = threadpool.makeRequests(spider, range(0,100,30))
[pool.putRequest(req) for req in requests]
pool.wait()
if name==‘main’:
# singelThread()
multiThread()

猜你喜欢

转载自blog.csdn.net/qq_29535381/article/details/89299505