不忘初心——你懂的!
学爬虫的初心是什么?大家的初心我不了解,而我只是为了以下这几行简单的代码。
(留存)
import requests from pyquery import PyQuery as py import time import os ##########由于域名不停变动,所以要获取最新域名############## url = 'https://www.ccc560.com' headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36'} print(url) response = requests.get(url,headers=headers) url = response.url ##########输入版块代码############## url_initial = url+'/htm/piclist1/'#Asia_photo ''' 想知道内容,尝试就可以了 girllist1~16 piclist1~4,6~8 ''' # s = requests.session() # proxies = { # "https": "http://118.254.147.216:3218" # } headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36'} print(url_initial) response = requests.get(url_initial,headers=headers) doc = py(response.text) ##########所需要的网页的部分网址都在某标签包围中############## ##########并对网页网址进行加工获取完整地址############## items = doc('li').items() for item in items: time.sleep(1) str1 = item.find('a').attr('href') url_pic = url+str1 print(url_pic) ##########为每一个网址都建立一个文件夹############## path = 'E:\\demo\\' + str1 if not os.path.exists(path): os.makedirs(path) ##########所需要的图片地址都在某标签的href属性中############## response_pic = requests.get(url_pic,headers=headers) doc_pic = py(response_pic.text) items_pic = doc_pic('div.picContent img').items() i = 0 for item in items_pic: str1 = '' pic = item.attr('src') print(pic) ##########为每一张图片赋予不重复的名字############## with open(path+'\\'+str(i)+'.jpg', 'wb')as f: cd = requests.get(pic).content f.write(cd) f.close() i=i + 1
只待江流汲海,万木朝东