1.前期准备
具体请查看上一篇
2.准备库requests,BeautifulSoup,xlwt,lxml
1.BeautifulSoup:是专业的网页爬取库,方便抓取网页信息
2.xlwt:生成excel表格
3.lxml:xml解析库
3.具体思路
企查查网站具有一定的反爬机制,直接爬取会受到网站阻拦,所以我们需要模拟浏览器请求,绕过反爬机制,打开企查查网站,获取cookie及一系列请求头文件,然后使用BeautifulSoup分析网页节点捕捉需要的信息
4.源码
# encoding: utf-8
import requests
from bs4 import BeautifulSoup
import lxml
import xlwt
import re
def craw():
file = xlwt.Workbook()
table = file.add_sheet('sheet1', cell_overwrite_ok=True)
print('正在爬取,请稍等....')
for n in range(1,500):
print('第'+ str(n) + '页......')
url = 'https://www.qichacha.com/g_JS_' + str(n) + '.html'
user_agent = 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36'
headers = {
'Host': 'www.qichacha.com',
'User-Agent': r'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36',
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8',
'Accept-Language': 'zh-CN,zh;q=0.9',
'Accept-Encoding': 'gzip, deflate, br',
'Referer': 'http://www.qichacha.com/',
'Cookie': r'zg_did=%7B%22did%22%3A%20%22166870cd07f60d-0c80294526eac7-36664c08-1fa400-166870cd0801af%22%7D; acw_tc=3af3b59815398640670163813ec3ddf30042b9b31607691a7b8d249c27; UM_distinctid=166870d292d85-016e1a972f471f-36664c08-1fa400-166870d292f349; _uab_collina=153986407973326937715323; QCCSESSID=g0gqbq7t1r8ksn94j8ii1qpbq1; CNZZDATA1254842228=364260894-1539860390-https%253A%252F%252Fwww.qichacha.com%252F%7C1540383468; Hm_lvt_3456bee468c83cc63fb5147f119f1075=1539864081,1540384169; zg_de1d1a35bfa24ce29bbf2c7eb17e6c4f=%7B%22sid%22%3A%201540384168992%2C%22updated%22%3A%201540384533698%2C%22info%22%3A%201539864055943%2C%22superProperty%22%3A%20%22%7B%7D%22%2C%22platform%22%3A%20%22%7B%7D%22%2C%22utm%22%3A%20%22%7B%7D%22%2C%22referrerDomain%22%3A%20%22%22%7D; Hm_lpvt_3456bee468c83cc63fb5147f119f1075=1540384534',
'Connection': 'keep-alive',
'If-Modified-Since': 'Wed, 24 Oct 2018 12:35:27 GMT',
'If-None-Match': '"59*******"',
'Cache-Control': 'private',
}
response = requests.get(url, headers=headers)
if response.status_code != 200:
response.encoding = 'utf-8'
print(response.status_code)
print('ERROR')
# soup = BeautifulSoup(response.text, 'lxml')
html_str = response.text
soup = BeautifulSoup(html_str, 'html.parser')
list = []
# list = soup.findAll(class_='panel panel-default')
list = soup.findAll(class_='panel panel-default')
for i in range(len(list)):
text = list[i]
soup2 = BeautifulSoup(str(text), 'lxml')
icon = soup2.find('img').attrs['src']
table.write((n - 1) * 10 + i, 1, str(icon))
name = soup2.find(class_='name').text
table.write((n - 1) * 10 + i, 2, name)
try:
content = soup2.findAll(class_='text-muted clear text-ellipsis m-t-xs')[0].text
address = soup2.findAll(class_='text-muted clear text-ellipsis m-t-xs')[1].text
table.write((n - 1) * 10 + i, 3, content)
table.write((n - 1) * 10 + i, 4, address)
except:
print('第'+str(n)+'页第'+str(i)+'行错误')
file.save('D:/qcc.xls')
if __name__ == '__main__':
craw()
5.结果
****6.申明
本次代码仅供学习使用,不应用于任何商业用途,若被查处,概不负责。
--完成