python爬虫7:完整的一个爬虫小demo

280次阅读
没有评论
python爬虫7:完整的一个爬虫小demo

from bs4 import BeautifulSoup import re import urllib.request import xlwt import sqlite3 import ssl

ssl._create_default_https_context = ssl._create_unverified_context def main(): #网址就不说了,避免侵权 baseUrl = "xxxxxx" data = getData(baseUrl) # saveDataToExcel(data) saveDataToDB(data)

def getData(baseUrl): headers = {"User-Agent": "Mozilla/5.0 (iPhone; CPU iPhone OS 13_2_3 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.0.3 Mobile/15E148 Safari/604.1", "cookie": "__51uvsct__JUTBIqXUbBvBL3hd=1; __51vcke__JUTBIqXUbBvBL3hd=f89fdbc5-1dbf-531a-8d48-608ad53241e3; __51vuft__JUTBIqXUbBvBL3hd=1639549481152; XSRF-TOKEN=eyJpdiI6ImFIK2VFbDg2NXQzbE9lTUZmNGFWMFE9PSIsInZhbHVlIjoiQ1NYU2dZVTV4NlpSM1FNOUVVdHFoZVB1N1daS2NoXC9hZUhOUThhcVZBbnpoYlBNTlwvZWtKUHNoNUdBdWxueTUwIiwibWFjIjoiMzMyMzE1MjRhMDU2MTIzMDBhM2UzZTQwMzJmMTk5YmM3NjgwMTJhMzM5MzZhZjcxM2I0YmRiZTYwNzBjNzM4ZCJ9; 91btlife_session=eyJpdiI6InEybEZiNUFUc3FwT1VMaGE1V3F4Qnc9PSIsInZhbHVlIjoiQ2M0UUY0ZFRDZGFQK3VEU3NyXC9VV1lsZ2dUZWQ2cHd2dVlMWVU4TktMYVZrN3RhTldPUXg4bmpub0ZSeHBmZkMwNHNweEd1Z09yMGQ1OTNld1dEYUdJY3FJOXZPTmpObUFcLzZxQjQ1MHl4Tlp6YW9SYXFBRzhGOGpLK1B6NnRBayIsIm1hYyI6ImVjYjM3OTFmNWI1NTMzMzEzM2Y0MzVmYzM1ZjNkNTNiMjM3ZmEyOTVlYmY5NjhiZDA5ODM5ZDkwNTkwM2RkMzgifQ%3D%3D; __vtins__JUTBIqXUbBvBL3hd=%7B%22sid%22%3A%20%222761075f-9fcf-5878-a827-2fecd1e51f8d%22%2C%20%22vd%22%3A%202%2C%20%22stt%22%3A%2083220%2C%20%22dr%22%3A%2083220%2C%20%22expires%22%3A%201639551364368%2C%20%22ct%22%3A%201639549564368%7D" } request = urllib.request.Request(url=baseUrl, headers=headers) rsp = urllib.request.urlopen(request)

soup = BeautifulSoup(rsp.read().decode('utf-8'), "html.parser") find_all = soup.find_all(class_="main-content") res={} titleList = {} detailList = [] i=0 #一级分类 titles = re.findall(r'style="margin-right: 7px;"></i>(.*?)</h4>', str(find_all[0])) for title in titles: title_temp={i:title} titleList.update(title_temp)

detail_htmls = re.findall('</i>'+title+'</h4>[sS]+?<h4 class="text-gray">', str(find_all[0])) if len(detail_htmls)>0: dts=re.findall('<strong>(.*?)</strong>',detail_htmls[0]) urls=re.findall(r'data-original-title="(.*?)"',detail_htmls[0]) for j in range(0, len(dts)): dlt = [] dlt.append(i) dlt.append(dts[j]) dlt.append(urls[j]) detailList.append(dlt) i+=1 res.update({"titleList":titleList}) res.update({"detailList":detailList}) return res

def saveDataToExcel(data): workbook = xlwt.Workbook(encoding="utf-8") sheet = workbook.add_sheet("sheet1") i=0 for title in data: sheet.write(i,0,title) i+=1 workbook.save("te.xls")

def saveDataToDB(data): connect = sqlite3.connect("test.db") c = connect.cursor() #获取游标 # i=1 # for title in data: # sql = r"insert into t1 (id,title) values (%d,'%s');"%(i,title) # i+=1 # print(sql) # c.execute(sql) titleList = data['titleList'] detailList = data['detailList'] for key,value in titleList.items(): sql = r"insert into title (id,title) values (%d,'%s');" %(key, value) c.execute(sql) for val in detailList: sql = r"insert into detail (`name`,titleId,url) values ('%s',%d,'%s');" %(val[1], val[0],val[2]) c.execute(sql) connect.commit() c.close()

if __name__ == "__main__": main()

神龙|纯净稳定代理IP免费测试>>>>>>>>天启|企业级代理IP免费测试>>>>>>>>IPIPGO|全球住宅代理IP免费测试

相关文章:

版权声明:Python教程2022-10-28发表,共计3013字。
新手QQ群:570568346,欢迎进群讨论 Python51学习