0%

如何利用python爬取网页图片-你懂的

如何利用python爬取网页图片

1.相关依赖库

  • BeautifulSoup 可以从HTML或XML文件中提取数据的Python库

  • requests Requests is an elegant and simple HTTP library for Python, built for human beings.

2.代码实现

代码编写调试过程,可以参考简书📕 小董不太懂的教程,其代码编写思路如下

1
2
3
4
5
6
7
st=>start: 开始
e=>end: 保存或处理爬去的信息
op1=>operation: 利用获取爬取的网页网址的html
op2=>operation: 利用BeautifulSoup将html信息转换成xml信息
op3=>operation: 利用BeautifulSoup挑选xml中想要的信息

st->op1->op2->op3->e

下面贴出代码全文

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
import requests
import os
from hashlib import md5
from requests.exceptions import RequestException
from bs4 import BeautifulSoup

headers = {'If-None-Match': 'W/"5cc2cd8f-2c58"',
"Referer": "http://www.mzitu.com/all/",
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/50.0.2661.102 UBrowser/6.1.2107.204 SafarMozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36'}
#请求头的这个Referer一定要加,妹子网有反爬,反正粘贴复制就行,多加几个信息无所谓
def get_page(url):
try:
response = requests.get(url, headers=headers)
if response.status_code == 200:
#print(response.text)
return response.text
return None
except RequestException:
print('获取索引页失败')
return None

def parse_page(html):
soup = BeautifulSoup(html, 'lxml')
items = soup.select('#pins li')
for link in items:
href = link.a['href']
#print(href)
get_detail_page(href)
# print(items)

def get_detail_page(href):
for i in range(1,100):
detail_url = href + '/' + str(i)
if requests.get(detail_url, headers=headers).status_code == 200:
#print(detail_url)
parse_detail_page(detail_url)
else:
print('已至末尾页')
return None


def parse_detail_page(detail_url):
try:
response = requests.get(detail_url, headers=headers)
if response.status_code == 200:
print('获取详情页成功')
detail_html = response.text
# print(detail_html)
get_image(detail_html)
return None
except RequestException:
print('获取详情页失败')
return None

def get_image(detail_html):
soup = BeautifulSoup(detail_html, 'lxml')
items= soup.select('.main-image')
# print(items)
for item in items:
#return item.img['src']
image = item.img['src']
save_image(image)

def save_image(image):
response = requests.get(image,headers=headers)
if response.status_code == 200:
data = response.content
file_path = '{0}/{1}.{2}'.format(os.getcwd(), md5(data).hexdigest(), 'jpg')
print(file_path)
if not os.path.exists(file_path):
with open(file_path, 'wb') as f:
f.write(data)
f.close()
print('保存成功')
else:
print('保存失败')

def main():
url = 'https://www.mzitu.com/tag/youhuo/'
html = get_page(url)
parse_page(html)


if __name__ == '__main__':
main()

[^1] https://www.jianshu.com/p/37202bec8f97

Donate comment here.

欢迎关注我的其它发布渠道