Python 简单网页爬虫
网上的妹子图爬虫:只爬取一个人物相册
import requests
from bs4 import BeautifulSoup
headers = {
'User-Agent':'Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1)',
'Referer':'http://www.mzitu.com'
}
# 初始链接
start_url = 'https://www.mzitu.com/161470'
start_html = requests.get(start_url,headers=headers) #生成一个response对象
# print(start_html.text) # .text是文本类型,如果是多媒体如音乐视频图片等,则是.content
soup = BeautifulSoup(start_html.content,'lxml')
max_span=soup.find('div',class_='pagenavi').find_all('span')[-2].get_text()
for page in range(1,int(max_span)+1):
page_url = start_url+'/'+str(page) #给初始链接加上页码数,就是某页的链接地址
image_page = requests.get(page_url,headers=headers)
# print(image_page.text)
image_soup = BeautifulSoup(image_page.content,'lxml')
image_url = image_soup.find('div',class_='main-image').find('img')['src'] #找到img标签的src属性的值,如<img src='lslsls'>,则返回lslsls
name = str(image_url) #别忘了转换类型
#print(name)
img = requests.get(name,headers = headers)
fpath = 'C:\\Users\\wztshine\\Desktop\\新建文件夹\\'+name[-7:] #对name参数切片,从倒数第七个开始。
with open(fpath, 'wb') as f:
print('output:', fpath)
f.write(img.content)