时间:2020-08-26 python爬虫 查看: 1042
实现思路:分为两部分,第一部分,获取网页上数据并使用xlwt生成excel(当然你也可以选择保存到数据库),第二部分获取网页数据使用IO流将图片保存到本地
一、爬取所有英雄属性并生成excel
1.代码
import json
import requests
import xlwt
# 设置头部信息,防止被检测出是爬虫
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.116 Safari/537.36'
}
url = "https://game.gtimg.cn/images/lol/act/img/js/heroList/hero_list.js"
response = requests.get(url=url, headers=headers).text
loads = json.loads(response)
dic = loads['hero']
# 创建一个workbook 设置编码
workbook = xlwt.Workbook(encoding='utf-8')
# 创建一个worksheet
worksheet = workbook.add_sheet('LOL')
try:
for i in range(len(dic)):
# 设置单元格宽度大小
worksheet.col(i).width = 4000
# 设置单元格高度大小
style = xlwt.easyxf('font:height 300')
worksheet.row(i).set_style(style)
# 第一行用于写入表头
if i == 0:
worksheet.write(i, 0, '编号')
worksheet.write(i, 1, '名称')
worksheet.write(i, 2, '英文名')
worksheet.write(i, 3, '中文名')
worksheet.write(i, 4, '角色')
worksheet.write(i, 5, '物攻')
worksheet.write(i, 6, '物防')
worksheet.write(i, 7, '魔攻')
worksheet.write(i, 8, '魔防')
continue
worksheet.write(i, 0, dic[i - 1]['heroId'])
worksheet.write(i, 1, dic[i - 1]['name'])
worksheet.write(i, 2, dic[i - 1]['alias'])
worksheet.write(i, 3, dic[i - 1]['title'])
roles_ = dic[i - 1]['roles']
roles = []
for data in roles_:
if 'mage' == data:
roles.append('法师')
if 'tank' == data:
roles.append('坦克')
if 'fighter' == data:
roles.append('战士')
if 'marksman' == data:
roles.append('ADC')
if 'assassin' == data:
roles.append('刺客')
if 'support' == data:
roles.append('辅助')
worksheet.write(i, 4, str(roles))
worksheet.write(i, 5, dic[i - 1]['attack'])
worksheet.write(i, 6, dic[i - 1]['defense'])
worksheet.write(i, 7, dic[i - 1]['magic'])
worksheet.write(i, 8, dic[i - 1]['difficulty'])
# 保存
workbook.save('C:\\Users\\Jonsson\\Desktop\\lol.xls')
except Exception as e:
print(e)
2.效果
二、爬取所有英雄皮肤并保存到本地
1.代码
import json
import os
import requests
import xlwt
# 设置头部信息,防止被检测出是爬虫
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.116 Safari/537.36'
}
url = "https://game.gtimg.cn/images/lol/act/img/js/heroList/hero_list.js"
# 请求英雄列表的url地址
response = requests.get(url=url, headers=headers).text
loads = json.loads(response)
dic = loads['hero']
for data in dic:
id_ = data['heroId']
skinUrl = 'https://game.gtimg.cn/images/lol/act/img/js/hero/%s.js' % id_
# 请求每个英雄皮肤的url地址
skinResponse = requests.get(url=skinUrl, headers=headers).text
json_loads = json.loads(skinResponse)
hero_ = json_loads['hero']
save_path = './skin/%s-%s-%s' % (hero_["heroId"], hero_['name'], hero_['title'])
# 文件夹不存在,则创建文件夹
folder = os.path.exists(save_path)
if not folder:
os.makedirs(save_path)
skins_ = json_loads['skins']
for data in skins_:
if data['chromas'] == '0':
content = requests.get(url=data['mainImg'], headers=headers).content
try:
with open('%s/%s.jpg' % (save_path, data['name']), "wb") as f:
print("正在下载英雄:%s 皮肤名称:%s ..." % (hero_['name'], data['name']))
f.write(content)
except Exception as e:
print('下载失败')
print(e)
2.效果
到此这篇关于用Python爬取LOL所有的英雄信息以及英雄皮肤的示例代码的文章就介绍到这了,更多相关Python爬取LOL所有英雄内容请搜索python博客以前的文章或继续浏览下面的相关文章希望大家以后多多支持python博客!