人妖在线一区,国产日韩欧美一区二区综合在线,国产啪精品视频网站免费,欧美内射深插日本少妇

新聞動(dòng)態(tài)

如何使用python爬取B站排行榜Top100的視頻數(shù)據(jù)

發(fā)布日期:2021-12-31 09:44 | 文章來源:CSDN

記得收藏呀?。?!

1、第三方庫導(dǎo)入

from bs4 import BeautifulSoup # 解析網(wǎng)頁
import re# 正則表達(dá)式,進(jìn)行文字匹配
import urllib.request,urllib.error  # 通過瀏覽器請求數(shù)據(jù)
import sqlite3  # 輕型數(shù)據(jù)庫
import time  # 獲取當(dāng)前時(shí)間

2、程序運(yùn)行主函數(shù)

爬取過程主要包括聲明爬取網(wǎng)頁 -> 爬取網(wǎng)頁數(shù)據(jù)并解析 -> 保存數(shù)據(jù)

def main():
	#聲明爬取網(wǎng)站
 baseurl = "https://www.bilibili.com/v/popular/rank/all"
 #爬取網(wǎng)頁
 datalist = getData(baseurl)
 # print(datalist)
 #保存數(shù)據(jù)
 dbname = time.strftime("%Y-%m-%d", time.localtime())
 dbpath = "BiliBiliTop100  " + dbname
 saveData(datalist,dbpath)

(1)在爬取的過程中采用的技術(shù)為:偽裝成瀏覽器對數(shù)據(jù)進(jìn)行請求;
(2)解析爬取到的網(wǎng)頁源碼時(shí):采用Beautifulsoup解析出需要的數(shù)據(jù),使用re正則表達(dá)式對數(shù)據(jù)進(jìn)行匹配;
(3)保存數(shù)據(jù)時(shí),考慮到B站排行榜是每日進(jìn)行刷新,故可以用當(dāng)前日期進(jìn)行保存數(shù)據(jù)庫命名。

3、程序運(yùn)行結(jié)果

數(shù)據(jù)庫中包含的數(shù)據(jù)有:排名、視頻鏈接、標(biāo)題、播放量、評(píng)論量、作者、綜合分?jǐn)?shù)這7個(gè)數(shù)據(jù)。

4、程序源代碼

from bs4 import BeautifulSoup #解析網(wǎng)頁
import re # 正則表達(dá)式,進(jìn)行文字匹配
import urllib.request,urllib.error
import sqlite3
import time

def main():
 #聲明爬取網(wǎng)站
 baseurl = "https://www.bilibili.com/v/popular/rank/all"
 #爬取網(wǎng)頁
 datalist = getData(baseurl)
 # print(datalist)
 #保存數(shù)據(jù)
 dbname = time.strftime("%Y-%m-%d", time.localtime())
 dbpath = "BiliBiliTop100  " + dbname
 saveData(datalist,dbpath)
#re正則表達(dá)式
findLink =re.compile(r'<a class="title" href="(.*?)" rel="external nofollow" ') #視頻鏈接
findOrder = re.compile(r'<div class="num">(.*?)</div>') #榜單次序
findTitle = re.compile(r'<a class="title" href=".*?" rel="external nofollow"  rel="external nofollow"  target="_blank">(.*?)</a>') #視頻標(biāo)題
findPlay = re.compile(r'<span class="data-box"><i class="b-icon play"></i>([\s\S]*)(.*?)</span> <span class="data-box">') #視頻播放量
findView = re.compile(r'<span class="data-box"><i class="b-icon view"></i>([\s\S]*)(.*?)</span> <a href=".*?" rel="external nofollow"  rel="external nofollow"  target="_blank"><span class="data-box up-name">') # 視頻評(píng)價(jià)數(shù)
findName = re.compile(r'<i class="b-icon author"></i>(.*?)</span></a>',re.S) #視頻作者
findScore = re.compile(r'<div class="pts"><div>(.*?)</div>綜合得分',re.S) #視頻得分
def getData(baseurl):
 datalist = []
 html = askURL(baseurl)
 #print(html)
 soup = BeautifulSoup(html,'html.parser')  #解釋器
 for item in soup.find_all('li',class_="rank-item"):
  # print(item)
  data = []
  item = str(item)
  Order = re.findall(findOrder,item)[0]
  data.append(Order)
  # print(Order)
  Link = re.findall(findLink,item)[0]
  Link = 'https:' + Link
  data.append(Link)
  # print(Link)
  Title = re.findall(findTitle,item)[0]
  data.append(Title)
  # print(Title)
  Play = re.findall(findPlay,item)[0][0]
  Play = Play.replace(" ","")
  Play = Play.replace("\n","")
  Play = Play.replace(".","")
  Play = Play.replace("萬","0000")
  data.append(Play)
  # print(Play)
  View = re.findall(findView,item)[0][0]
  View = View.replace(" ","")
  View = View.replace("\n","")
  View = View.replace(".","")
  View = View.replace("萬","0000")
  data.append(View)
  # print(View)
  Name = re.findall(findName,item)[0]
  Name = Name.replace(" ","")
  Name = Name.replace("\n","")
  data.append(Name)
  # print(Name)
  Score = re.findall(findScore,item)[0]
  data.append(Score)
  # print(Score)
  datalist.append(data)
 return datalist
def askURL(url):
 #設(shè)置請求頭
 head = {
  "User-Agent": "Mozilla/5.0 (Windows NT 10.0;Win64;x64) AppleWebKit/537.36(KHTML, likeGecko) Chrome/80.0.3987.163Safari/537.36"
 }
 request = urllib.request.Request(url, headers = head)
 html = ""
 try:
  response = urllib.request.urlopen(request)
  html = response.read().decode("utf-8")
  #print(html)
 except urllib.error.URLError as e:
  if hasattr(e,"code"):
print(e.code)
  if hasattr(e,"reason"):
print(e.reason)
 return html
def saveData(datalist,dbpath):
 init_db(dbpath)
 conn = sqlite3.connect(dbpath)
 cur = conn.cursor()
 for data in datalist:
  sql = '''
  insert into Top100(
  id,info_link,title,play,view,name,score)
  values("%s","%s","%s","%s","%s","%s","%s")'''%(data[0],data[1],data[2],data[3],data[4],data[5],data[6])
  print(sql)
  cur.execute(sql)
  conn.commit()
 cur.close()
 conn.close()
def init_db(dbpath):
 sql = '''
 create table Top100
 (
 id integer primary key autoincrement,
 info_link text,
 title text,
 play numeric,
 view numeric,
 name text,
 score numeric
 )
 '''
 conn = sqlite3.connect(dbpath)
 cursor = conn.cursor()
 cursor.execute(sql)
 conn.commit()
 conn.close()

if __name__ =="__main__":
 main()

到此這篇關(guān)于如何使用python爬取B站排行榜Top100的視頻數(shù)據(jù)的文章就介紹到這了,更多相關(guān)python B站視頻 內(nèi)容請搜索本站以前的文章或繼續(xù)瀏覽下面的相關(guān)文章希望大家以后多多支持本站!

版權(quán)聲明:本站文章來源標(biāo)注為YINGSOO的內(nèi)容版權(quán)均為本站所有,歡迎引用、轉(zhuǎn)載,請保持原文完整并注明來源及原文鏈接。禁止復(fù)制或仿造本網(wǎng)站,禁止在非www.sddonglingsh.com所屬的服務(wù)器上建立鏡像,否則將依法追究法律責(zé)任。本站部分內(nèi)容來源于網(wǎng)友推薦、互聯(lián)網(wǎng)收集整理而來,僅供學(xué)習(xí)參考,不代表本站立場,如有內(nèi)容涉嫌侵權(quán),請聯(lián)系alex-e#qq.com處理。

相關(guān)文章

實(shí)時(shí)開通

自選配置、實(shí)時(shí)開通

免備案

全球線路精選!

全天候客戶服務(wù)

7x24全年不間斷在線

專屬顧問服務(wù)

1對1客戶咨詢顧問

在線
客服

在線客服:7*24小時(shí)在線

客服
熱線

400-630-3752
7*24小時(shí)客服服務(wù)熱線

關(guān)注
微信

關(guān)注官方微信
頂部