花瓣图片的加载使用了延迟加载的技术,源代码只能下载20多张图片,修改后基本能下载所有的了,只是速度有点慢,后面再优化下
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
|
import urllib, urllib2, re, sys, os,requests path = r "C:\wqa\beautify" url = 'http://huaban.com/favorite/beauty' #http://huaban.com/explore/zhongwenlogo/?ig1un9tq&max=327773629&limit=20&wfl=1 i_headers = { "User-Agent" : "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/46.0.2490.71 Safari/537.36" } count = 0 def urlHandle(url): req = urllib2.Request(url, headers = i_headers) html = urllib2.urlopen(req).read() reg = re. compile (r '"pin_id":(\d+),.+?"file":{"farm":"farm1", "bucket":"hbimg",.+?"key":"(.*?)",.+?"type":"image/(.*?)"' , re.S) groups = re.findall(reg, html) return groups def imgHandle(groups): if groups: for att in groups: pin_id = att[ 0 ] att_url = att[ 1 ] + '_fw236' img_type = att[ 2 ] img_url = 'http://img.hb.aicdn.com/' + att_url r = requests.get(img_url) with open (path + att_url + '.' + img_type, 'wb' ) as fd: for chunk in r.iter_content(): fd.write(chunk) groups = urlHandle(url) imgHandle(groups) while (groups): count + = 1 print count pin_id = groups[ - 1 ][ 0 ] print pin_id urltemp = url + '/?max=' + str (pin_id) + '&limit=' + str ( 20 ) + '&wfl=1' print (urltemp) groups = urlHandle(urltemp) #print groups imgHandle(groups) |