企查查简单爬虫
Posted meipu
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了企查查简单爬虫相关的知识,希望对你有一定的参考价值。
经历过企查查这个网站后,强烈感觉到使用抓包的重要性,以至于决定从此以后使用抓包进行模拟请求,放弃使用F12进行分析。
写下这篇文章,奠基死去的F12~~~
1 import requests 2 from lxml import etree 3 4 url = "https://www.qcc.com/search?key=%E5%A4%A9%E6%B4%A5%E6%BB%A8%E6%B5%B7%E6%96%B0%E5%8C%BA" 5 6 hed = { 7 "host": "www.qcc.com", 8 "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (Khtml, like Gecko) Chrome/83.0.4103.106 Safari/537.36", 9 "upgrade-insecure-requests": "1", 10 "cookie": "QCCSESSID=vpk1mpc45ci95eu83etg528881; zg_did=%7B%22did%22%3A%20%221732cdcac86bf-0039dd6baef69a-4353761-100200-1732cdcac8844f%22%7D; UM_distinctid=1732cdcb0a713b-01b058b949aa5a-4353761-100200-1732cdcb0ab44e; hasShow=1; _uab_collina=159418552807339394444789; acw_tc=7d27c71c15941953776602556e6b8442bc8001e4e1270e8fead4b79557; CNZZDATA1254842228=1092104090-1594185078-https%253A%252F%252Fwww.baidu.com%252F%7C1594195878; Hm_lvt_78f134d5a9ac3f92524914d0247e70cb=1594194111,1594195892,1594195918,1594196042; Hm_lpvt_78f134d5a9ac3f92524914d0247e70cb=1594196294; zg_de1d1a35bfa24ce29bbf2c7eb17e6c4f=%7B%22sid%22%3A%201594185526424%2C%22updated%22%3A%201594196294349%2C%22info%22%3A%201594185526455%2C%22superProperty%22%3A%20%22%7B%7D%22%2C%22platform%22%3A%20%22%7B%7D%22%2C%22utm%22%3A%20%22%7B%5C%22%24utm_source%5C%22%3A%20%5C%22baidu1%5C%22%2C%5C%22%24utm_medium%5C%22%3A%20%5C%22cpc%5C%22%2C%5C%22%24utm_term%5C%22%3A%20%5C%22pzsy%5C%22%7D%22%2C%22referrerDomain%22%3A%20%22www.baidu.com%22%2C%22cuid%22%3A%20%22fd05f1ac2b561244aaa6b27b3bb617a4%22%7D", 11 } 12 13 resq = requests.get(url = url,headers = hed).content 14 response = etree.HTML(resq) 15 16 title_list = [] 17 title = response.xpath(‘//*[@id="search-result"]//tr/td[3]/a//text()‘) 18 for tit in title: 19 tit = tit.replace(‘,‘,‘‘).strip() 20 title_list.append(tit) 21 22 addr_list = [] 23 addrs = response.xpath(‘//*[@id="search-result"]//tr/td[3]/p[4]//text()‘) 24 for addr in addrs: 25 addr = addr.replace(‘,‘,‘‘).strip() 26 addr_list.append(addr) 27 28 print(title_list) 29 print(addr_list)
代码很简单,甚至于简陋,为什么要记录下这个爬虫,因为请求头部信息,自己进行分析,和ctrl+c+v导致请求头数据不准确,严重感觉到抓包工具的请求分析更加快速有效。
继续加油,继续努力
以上是关于企查查简单爬虫的主要内容,如果未能解决你的问题,请参考以下文章