目标网址 https://www.csflhjw.com/zhenghun/34.html?page=1
一、打开界面
鼠标右键打开检查,方框里为你一个文小姐的征婚信息。。由此判断出为同步加载
点击elements,定位图片地址,方框里为该女士的url地址及图片地址
可以看出该女士的url地址不全,之后在代码中要进行url的拼接,看一下翻页的url地址有什么变化
点击第2页
https://www.csflhjw.com/zhenghun/34.html?page=2
点击第3页
https://www.csflhjw.com/zhenghun/34.html?page=3
可以看出变化在最后
做一下fou循环格式化输出一下。。一共10页
二、代码解析
1.获取所有的女士的url,xpath的路径就不详细说了。。
2.构造每一位女士的url地址
3.然后点开一位女士的url地址,用同样的方法,确定也为同步加载
4.之后就是女士url地址html的xpath提取,每个都打印一下,把不要的过滤一下
5.最后就是文件的保存
打印结果:
三、完整代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
|
# !/usr/bin/nev python # -*-coding:utf8-*- import requests, os, csv from pprint import pprint from lxml import etree def main(): for i in range ( 1 , 11 ): start_url = 'https://www.csflhjw.com/zhenghun/34.html?page={}' . format (i) headers = { 'user-agent' : 'mozilla/5.0 (windows nt 10.0; wow64) applewebkit/537.36 (khtml, like gecko) ' 'chrome/87.0.4280.88 safari/537.36' } response = requests.get(start_url, headers = headers).content.decode() # # pprint(response) # 3 解析数据 html_str = etree.html(response) info_urls = html_str.xpath(r '//div[@class="e"]/div[@class="e-img"]/a/@href' ) # pprint(info_urls) # 4、循环遍历 构造img_info_url for info_url in info_urls: info_url = r 'https://www.csflhjw.com' + info_url # print(info_url) # 5、对info_url发请求,解析得到img_urls response = requests.get(info_url, headers = headers).content.decode() html_str = etree.html(response) # pprint(html_str) img_url = 'https://www.csflhjw.com/' + html_str.xpath(r '/html/body/div[4]/div/div[1]/div[2]/div[1]/div[' r '1]/img/@src' )[ 0 ] # pprint(img_url) name = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/h2/text()' )[ 0 ] # pprint(name) xueli = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[1]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(xueli) job = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[2]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(job) marital_status = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[3]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(marital_status) is_child = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[4]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(is_child) home = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[5]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(home) workplace = html_str.xpath(r '//div[@class="team-info"]/div[@class="team-e"]/p[6]/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(workplace) requ = html_str.xpath(r '/html/body/div[4]/div/div[1]/div[2]/div[2]/div[2]/p[2]/span/text()' )[ 0 ].split( ':' )[ 1 ] # pprint(requ) requ = [requ if requ ! = str () else '无要求' ][ 0 ] monologue = html_str.xpath(r '//div[@class="hunyin-1-3"]/p/text()' ) # pprint(monologue) monologue = [monologue[ 0 ].replace( ' ' , ' ').replace(' \xa0 ', ' ') if monologue !=list() else ' 无'][ 0 ] # pprint(monologue) zeo_age = html_str.xpath(r '/html/body/div[4]/div/div[1]/div[2]/div[2]/div[2]/p[1]/span[1]/text()' )[ 0 ].split( ':' )[ 1 ] zeo_age = [zeo_age if zeo_age! = str () else '无要求' ][ 0 ] # pprint(zeo_age) zeo_address = html_str.xpath(r '/html/body/div[4]/div/div[1]/div[2]/div[2]/div[2]/p[1]/span[2]/text()' )[ 0 ].split( ':' )[ 1 ] zeo_address = [zeo_address if zeo_address! = str () else '无要求' ][ 0 ] # pprint(zeo_address) if not os.path.exists(r './{}' . format ( '妹子信息数据' )): os.mkdir(r './{}' . format ( '妹子信息数据' )) csv_header = [ '姓名' , '学历' , '职业' , '婚姻状况' , '有无子女' , '是否购房' , '工作地点' , '择偶年龄' , '择偶城市' , '择偶要求' , '个人独白' , '照片链接' ] with open (r './{}/{}.csv' . format ( '妹子信息数据' , '妹子数据' ), 'w' , newline = ' ', encoding=' gbk') as file_csv: csv_writer_header = csv.dictwriter(file_csv, csv_header) csv_writer_header.writeheader() try : with open (r './{}/{}.csv' . format ( '妹子信息数据' , '妹子数据' ), 'a+' , newline = '', encoding = 'gbk' ) as file_csv: csv_writer = csv.writer(file_csv, delimiter = ',' ) csv_writer.writerow([name, xueli, job, marital_status, is_child, home, workplace, zeo_age, zeo_address, requ, monologue, img_url]) print (r '***妹子信息数据:{}' . format (name)) except exception as e: with open (r './{}/{}.csv' . format ( '妹子信息数据' , '妹子数据' ), 'a+' , newline = '', encoding = 'utf-8' ) as file_csv: csv_writer = csv.writer(file_csv, delimiter = ',' ) csv_writer.writerow([name, xueli, job, marital_status, is_child, home, workplace, zeo_age, zeo_address, requ, monologue, img_url]) print (r '***妹子信息数据保存成功:{}' . format (name)) if __name__ = = '__main__' : main() |
到此这篇关于单身狗福利?python爬取某婚恋网征婚数据的文章就介绍到这了,更多相关python爬取征婚数据内容请搜索服务器之家以前的文章或继续浏览下面的相关文章希望大家以后多多支持服务器之家!
原文链接:https://blog.csdn.net/weixin_54733110/article/details/117360958