Python | 用Python爬取LOL所有的英雄信息以及英雄皮肤

实现思路:分为两部分,第一部分,获取网页上数据并使用xlwt生成excel(当然你也可以选择保存到数据库),第二部分获取网页数据使用IO流将图片保存到本地

一、爬取所有英雄属性并生成excel

1.代码

import json

import requests
import xlwt

# 设置头部信息,防止被检测出是爬虫
headers = {
    
    
    'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.116 Safari/537.36'
}
url = "https://game.gtimg.cn/images/lol/act/img/js/heroList/hero_list.js"
response = requests.get(url=url, headers=headers).text
loads = json.loads(response)
dic = loads['hero']
# 创建一个workbook 设置编码
workbook = xlwt.Workbook(encoding='utf-8')
# 创建一个worksheet
worksheet = workbook.add_sheet('LOL')
try:
    for i in range(len(dic)):
        # 设置单元格宽度大小
        worksheet.col(i).width = 4000
        # 设置单元格高度大小
        style = xlwt.easyxf('font:height 300')
        worksheet.row(i).set_style(style)
        # 第一行用于写入表头
        if i == 0:
            worksheet.write(i, 0, '编号')
            worksheet.write(i, 1, '名称')
            worksheet.write(i, 2, '英文名')
            worksheet.write(i, 3, '中文名')
            worksheet.write(i, 4, '角色')
            worksheet.write(i, 5, '物攻')
            worksheet.write(i, 6, '物防')
            worksheet.write(i, 7, '魔攻')
            worksheet.write(i, 8, '魔防')
            continue
        worksheet.write(i, 0, dic[i - 1]['heroId'])
        worksheet.write(i, 1, dic[i - 1]['name'])
        worksheet.write(i, 2, dic[i - 1]['alias'])
        worksheet.write(i, 3, dic[i - 1]['title'])
        roles_ = dic[i - 1]['roles']
        roles = []
        for data in roles_:
            if 'mage' == data:
                roles.append('法师')
            if 'tank' == data:
                roles.append('坦克')
            if 'fighter' == data:
                roles.append('战士')
            if 'marksman' == data:
                roles.append('ADC')
            if 'assassin' == data:
                roles.append('刺客')
            if 'support' == data:
                roles.append('辅助')
        worksheet.write(i, 4, str(roles))
        worksheet.write(i, 5, dic[i - 1]['attack'])
        worksheet.write(i, 6, dic[i - 1]['defense'])
        worksheet.write(i, 7, dic[i - 1]['magic'])
        worksheet.write(i, 8, dic[i - 1]['difficulty'])
        # 保存
    workbook.save('C:\\Users\\Jonsson\\Desktop\\lol.xls')
except Exception as e:
    print(e)

2.效果

在这里插入图片描述

二、爬取所有英雄皮肤并保存到本地

1.代码

import json
import os

import requests
import xlwt

# 设置头部信息,防止被检测出是爬虫
headers = {
    
    
    'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.116 Safari/537.36'
}
url = "https://game.gtimg.cn/images/lol/act/img/js/heroList/hero_list.js"
# 请求英雄列表的url地址
response = requests.get(url=url, headers=headers).text
loads = json.loads(response)
dic = loads['hero']
for data in dic:
    id_ = data['heroId']
    skinUrl = 'https://game.gtimg.cn/images/lol/act/img/js/hero/%s.js' % id_
    # 请求每个英雄皮肤的url地址
    skinResponse = requests.get(url=skinUrl, headers=headers).text
    json_loads = json.loads(skinResponse)
    hero_ = json_loads['hero']
    save_path = './skin/%s-%s-%s' % (hero_["heroId"], hero_['name'], hero_['title'])
    # 文件夹不存在,则创建文件夹
    folder = os.path.exists(save_path)
    if not folder:
        os.makedirs(save_path)
    skins_ = json_loads['skins']
    for data in skins_:
        if data['chromas'] == '0':
            content = requests.get(url=data['mainImg'], headers=headers).content
            try:
                with open('%s/%s.jpg' % (save_path, data['name']), "wb") as f:
                    print("正在下载英雄:%s 皮肤名称:%s ..." % (hero_['name'], data['name']))
                    f.write(content)
            except Exception as e:
                print('下载失败')
                print(e)

2.效果

在这里插入图片描述
在这里插入图片描述
点赞收藏关注,你的支持是我最大的动力!

猜你喜欢

转载自blog.csdn.net/y1534414425/article/details/107247322