如何使用python爬取知乎熱榜Top50數(shù)據(jù)
1、導(dǎo)入第三方庫(kù)
import urllib.request,urllib.error #請(qǐng)求網(wǎng)頁(yè) from bs4 import BeautifulSoup # 解析數(shù)據(jù) import sqlite3 # 導(dǎo)入數(shù)據(jù)庫(kù) import re # 正則表達(dá)式 import time # 獲取當(dāng)前時(shí)間
2、程序的主函數(shù)
def main():
# 聲明爬取網(wǎng)頁(yè)
baseurl = "https://www.zhihu.com/hot"
# 爬取網(wǎng)頁(yè)
datalist = getData(baseurl)
#保存數(shù)據(jù)
dbname = time.strftime("%Y-%m-%d", time.localtime()) #
dbpath = "zhihuTop50 " + dbname
saveData(datalist,dbpath)
3、正則表達(dá)式匹配數(shù)據(jù)
#正則表達(dá)式 findlink = re.compile(r'<a class="css-hi1lih" href="(.*?)" rel="external nofollow" rel="external nofollow" ') #問(wèn)題鏈接 findid = re.compile(r'<div class="css-blkmyu">(.*?)</div>') #問(wèn)題排名 findtitle = re.compile(r'<h1 class="css-3yucnr">(.*?)</h1>') #問(wèn)題標(biāo)題 findintroduce = re.compile(r'<div class="css-1o6sw4j">(.*?)</div>') #簡(jiǎn)要介紹 findscore = re.compile(r'<div class="css-1iqwfle">(.*?)</div>') #熱門評(píng)分 findimg = re.compile(r'<img class="css-uw6cz9" src="(.*?)"/>') #文章配圖
4、程序運(yùn)行結(jié)果


5、程序源代碼
import urllib.request,urllib.error
from bs4 import BeautifulSoup
import sqlite3
import re
import time
def main():
# 聲明爬取網(wǎng)頁(yè)
baseurl = "https://www.zhihu.com/hot"
# 爬取網(wǎng)頁(yè)
datalist = getData(baseurl)
#保存數(shù)據(jù)
dbname = time.strftime("%Y-%m-%d", time.localtime())
dbpath = "zhihuTop50 " + dbname
saveData(datalist,dbpath)
print()
#正則表達(dá)式
findlink = re.compile(r'<a class="css-hi1lih" href="(.*?)" rel="external nofollow" rel="external nofollow" ') #問(wèn)題鏈接
findid = re.compile(r'<div class="css-blkmyu">(.*?)</div>') #問(wèn)題排名
findtitle = re.compile(r'<h1 class="css-3yucnr">(.*?)</h1>') #問(wèn)題標(biāo)題
findintroduce = re.compile(r'<div class="css-1o6sw4j">(.*?)</div>') #簡(jiǎn)要介紹
findscore = re.compile(r'<div class="css-1iqwfle">(.*?)</div>') #熱門評(píng)分
findimg = re.compile(r'<img class="css-uw6cz9" src="(.*?)"/>') #文章配圖
def getData(baseurl):
datalist = []
html = askURL(baseurl)
# print(html)
soup = BeautifulSoup(html,'html.parser')
for item in soup.find_all('a',class_="css-hi1lih"):
# print(item)
data = []
item = str(item)
Id = re.findall(findid,item)
if(len(Id) == 0):
Id = re.findall(r'<div class="css-mm8qdi">(.*?)</div>',item)[0]
else: Id = Id[0]
data.append(Id)
# print(Id)
Link = re.findall(findlink,item)[0]
data.append(Link)
# print(Link)
Title = re.findall(findtitle,item)[0]
data.append(Title)
# print(Title)
Introduce = re.findall(findintroduce,item)
if(len(Introduce) == 0):
Introduce = " "
else:Introduce = Introduce[0]
data.append(Introduce)
# print(Introduce)
Score = re.findall(findscore,item)[0]
data.append(Score)
# print(Score)
Img = re.findall(findimg,item)
if (len(Img) == 0):
Img = " "
else: Img = Img[0]
data.append(Img)
# print(Img)
datalist.append(data)
return datalist
def askURL(baseurl):
# 設(shè)置請(qǐng)求頭
head = {
# "User-Agent": "Mozilla/5.0 (Windows NT 10.0;Win64;x64) AppleWebKit/537.36(KHTML, likeGecko) Chrome/80.0.3987.163Safari/537.36"
"User-Agent": "Mozilla / 5.0(iPhone;CPUiPhoneOS13_2_3likeMacOSX) AppleWebKit / 605.1.15(KHTML, likeGecko) Version / 13.0.3Mobile / 15E148Safari / 604.1"
}
request = urllib.request.Request(baseurl, headers=head)
html = ""
try:
response = urllib.request.urlopen(request)
html = response.read().decode("utf-8")
# print(html)
except urllib.error.URLError as e:
if hasattr(e, "code"):
print(e.code)
if hasattr(e, "reason"):
print(e.reason)
return html
print()
def saveData(datalist,dbpath):
init_db(dbpath)
conn = sqlite3.connect(dbpath)
cur = conn.cursor()
for data in datalist:
sql = '''
insert into Top50(
id,info_link,title,introduce,score,img)
values("%s","%s","%s","%s","%s","%s")'''%(data[0],data[1],data[2],data[3],data[4],data[5])
print(sql)
cur.execute(sql)
conn.commit()
cur.close()
conn.close()
def init_db(dbpath):
sql = '''
create table Top50
(
id integer primary key autoincrement,
info_link text,
title text,
introduce text,
score text,
img text
)
'''
conn = sqlite3.connect(dbpath)
cursor = conn.cursor()
cursor.execute(sql)
conn.commit()
conn.close()
if __name__ =="__main__":
main()
到此這篇關(guān)于如何使用python爬取知乎熱榜Top50數(shù)據(jù)的文章就介紹到這了,更多相關(guān)python 爬取知乎內(nèi)容請(qǐng)搜索腳本之家以前的文章或繼續(xù)瀏覽下面的相關(guān)文章希望大家以后多多支持腳本之家!
相關(guān)文章
如何使用?Python為你的在線會(huì)議創(chuàng)建一個(gè)假的攝像頭
這篇文章主要介紹了使用?Python為你的在線會(huì)議創(chuàng)建一個(gè)假的攝像頭,在?Python?的幫助下,不再?gòu)?qiáng)制開(kāi)啟攝像頭,將向你展示如何為你的在線會(huì)議創(chuàng)建一個(gè)假的攝像頭,需要的朋友可以參考下2022-08-08
Python 實(shí)現(xiàn)局域網(wǎng)遠(yuǎn)程屏幕截圖案例
這篇文章主要介紹了Python 實(shí)現(xiàn)局域網(wǎng)遠(yuǎn)程屏幕截圖案例,具有很好的參考價(jià)值,希望對(duì)大家有所幫助。一起跟隨小編過(guò)來(lái)看看吧2021-03-03
Django框架cookie和session方法及參數(shù)設(shè)置
這篇文章主要為大家介紹了Django框架cookie和session參數(shù)設(shè)置及介紹,有需要的朋友可以借鑒參考下,希望能夠有所幫助,祝大家多多進(jìn)步,早日升職加薪2022-03-03
利用Python和C++實(shí)現(xiàn)解析gltf文件
gltf是類似于stl、obj、ply等常見(jiàn)的3D對(duì)象存儲(chǔ)格式,它被設(shè)計(jì)出來(lái)是為了便于渲染的數(shù)據(jù)轉(zhuǎn)換和傳輸,本文為大家介紹了使用Python和C++解析gltf文件的方法,感興趣的可以了解下2023-09-09
Python項(xiàng)目實(shí)戰(zhàn)之使用Django框架實(shí)現(xiàn)支付寶付款功能
這篇文章主要介紹了Python項(xiàng)目實(shí)戰(zhàn)之使用Django框架實(shí)現(xiàn)支付寶付款功能,本文給大家介紹的非常詳細(xì),對(duì)大家的學(xué)習(xí)或工作具有一定的參考借鑒價(jià)值,需要的朋友可以參考下2021-02-02
Python實(shí)現(xiàn)控制手機(jī)電腦拍照并自動(dòng)發(fā)送郵箱
這篇文章主要介紹了如何實(shí)現(xiàn)利用Python控制手機(jī)電腦拍照并自動(dòng)發(fā)送郵箱,文中的示例代碼講解詳細(xì),感興趣的小伙伴可以跟隨小編一起動(dòng)手試一試2022-02-02

