天天看点

python 爬取 mm131 图片

#!/usr/bin/env python
#-*-conding:utf-8-*-

import requests # 发送http请求
from bs4 import BeautifulSoup # 解析html
import lxml # 解析器 中文不乱码
import os #创建文件夹

# 爬取 mm131 图片
start_url = "http://www.mm131.com/xinggan/"

headers = {
'User-Agent':'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.26 Safari/537.36 Core/1.63.6788.400 QQBrowser/10.3.2864.400',
'Referer':start_url
}
path = 'C:/Users/Administrator/Desktop/tmp' #保存文件路径
folder = '/mm131/'

# 下载图片
def download_img(img_url, path):
    img = requests.get(img_url,  headers=headers)
    name = path+'/'+img_url.split('/')[-1]
    with open(name, "ab") as f:
        f.write(img.content)


# 1.获取html内容
ret_html = requests.get(url=start_url, headers=headers)


# 2.解析html内容
soup = BeautifulSoup(ret_html.content, "lxml")

# 查找标签
a_list = soup.find("div", attrs={"class":"main"}).find("dl",attrs={"class":"list-left public-box"}).find_all("a",  attrs={"target":"_blank"}, recursive=True)

url_list = []
for i in a_list:
    url_list.append(i["href"])


for i in url_list:

    headers['Referer'] = i
    ret_html = requests.get(url=i, headers=headers)
    soup = BeautifulSoup(ret_html.content, "html.parser")

    title = soup.find("h5").text # 文件名
    page = soup.find("div",attrs={'class':'content-page'}).find_all('a')[-2].text
    img_url = soup.find("div",attrs={'class':'content-pic'}).find("img")['src']
    
    if not os.path.isdir(path+folder+title):
        os.makedirs(path+folder+title)

    # 下载第一张图片
    headers['Referer'] = i
    download_img(img_url, path+folder+title)

    for j in range(2,int(page)+1):
        img_url2 = img_url.split('1.jpg')[0]+str(j)+'.jpg'
        # 下载图片
        headers['Referer'] = i.split('.html')[0]+'_'+i+'.html'
        download_img(img_url2, path+folder+title)