????簡(jiǎn)單記錄一下爬取網(wǎng)站圖片保存到本地指定目錄過程,希望對(duì)剛?cè)腴T的小伙伴有所幫助!
????目標(biāo)網(wǎng)站就是下圖所示頁面:
實(shí)現(xiàn)步驟:
????1.爬取每頁的圖片地址集合
????2.下載圖片到本地
????3. 獲取指定頁數(shù)的頁面路徑
以下是實(shí)現(xiàn)代碼:文章來源地址http://www.zghlxwxcb.cn/news/detail-733704.html
import bs4
import requests
import os
# 下載圖片到本地
def down_load_img(local_path,img_url):
img_name = os.path.basename(img_url[img_url.rfind("/")+1:len(img_url)])
with open(f"{local_path}\{img_name}","wb") as imgFile:
res = requests.get(img_url)
if res.status_code == 200:
resp = imgFile.write(res.content)
if resp is not None:
print(f"{img_name}下載成功!")
else:
print(f"{img_name}下載失敗!")
else:
print(f"{img_name}請(qǐng)求失敗,下載失敗!")
# 獲取每頁的圖片地址
def get_img_url(website_url):
return_img_url_list=[]
res = requests.get(website_url)
if res.status_code != 200:
print("請(qǐng)求失敗!")
website_content = res.content
soup = bs4.BeautifulSoup(website_content, "html.parser")
img_url_list = soup.find_all("div", class_="pic")
for imgUrl in img_url_list:
style_info = imgUrl["style"]
img = style_info[style_info.find("(") + 1:style_info.find(")")]
return_img_url_list.append(img)
return return_img_url_list
# 獲取指定頁數(shù)的頁面路徑
def get_website_url(page_num):
website_format="https://pic.netbian.top/4kmeinv/index_{}.html"
web_site_url_list=[]
for i in range(1,page_num+1):
web_site_url_list.append(f"https://pic.netbian.top/4kmeinv/index_{i}.html")
return web_site_url_list
if __name__ == '__main__':
local_path="D:\mvImg"
page_num=2
for website_url in get_website_url(page_num):
for img_url in get_img_url(website_url):
down_load_img(local_path,img_url)
文章來源:http://www.zghlxwxcb.cn/news/detail-733704.html
到了這里,關(guān)于python入門實(shí)戰(zhàn):爬取圖片到本地的文章就介紹完了。如果您還想了解更多內(nèi)容,請(qǐng)?jiān)谟疑辖撬阉鱐OY模板網(wǎng)以前的文章或繼續(xù)瀏覽下面的相關(guān)文章,希望大家以后多多支持TOY模板網(wǎng)!