欢迎您访问程序员文章站本站旨在为大家提供分享程序员计算机编程知识!
您现在的位置是: 首页

不忘初心——你懂的!

程序员文章站 2022-06-17 18:25:01
...

学爬虫的初心是什么?大家的初心我不了解,而我只是为了以下这几行简单的代码。

(留存)

 

import requests
from pyquery import PyQuery as py
import time
import os
##########由于域名不停变动,所以要获取最新域名##############
url = 'https://www.ccc560.com'
headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36'}
print(url)
response = requests.get(url,headers=headers)
url = response.url

##########输入版块代码##############

url_initial = url+'/htm/piclist1/'#Asia_photo
'''
想知道内容,尝试就可以了
girllist1~16
piclist1~4,6~8
'''
# s = requests.session()
# proxies = {
#   "https": "http://118.254.147.216:3218"
# }
headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36'}
print(url_initial)
response = requests.get(url_initial,headers=headers)
doc = py(response.text)

##########所需要的网页的部分网址都在某标签包围中##############
##########并对网页网址进行加工获取完整地址##############
items = doc('li').items()
for item in items:
    time.sleep(1)
    str1 = item.find('a').attr('href')
    url_pic = url+str1
    print(url_pic)

    ##########为每一个网址都建立一个文件夹##############
    path = 'E:\\demo\\' + str1
    if not os.path.exists(path):
        os.makedirs(path)

    ##########所需要的图片地址都在某标签的href属性中##############
    response_pic = requests.get(url_pic,headers=headers)
    doc_pic = py(response_pic.text)
    items_pic = doc_pic('div.picContent img').items()
    i = 0
    for item in items_pic:
        str1 = ''
        pic = item.attr('src')
        print(pic)

        ##########为每一张图片赋予不重复的名字##############
        with open(path+'\\'+str(i)+'.jpg', 'wb')as f:
            cd = requests.get(pic).content
            f.write(cd)
            f.close()
        i=i + 1