本文实例讲述了Python实现的爬取百度文库功能。分享给大家供大家参考,具体如下:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
|
# -*- coding: utf-8 -*- from selenium import webdriver from bs4 import BeautifulSoup from docx import Document from docx.enum.text import WD_ALIGN_PARAGRAPH # 用来居中显示标题 from time import sleep from selenium.webdriver.common.keys import Keys # 浏览器安装路径 #BROWSER_PATH=\'C:\Users\Administrator\AppData\Local\Google\Chrome\Application\chromedriver.exe' #目的URL DEST_URL = 'https://wenku.baidu.com/view/aa31a84bcf84b9d528ea7a2c.html' #用来保存文档 doc_title = '' doc_content_list = [] def find_doc(driver, init = True ): global doc_content_list global doc_title stop_condition = False html = driver.page_source soup1 = BeautifulSoup(html, 'lxml' ) if (init is True ): # 得到标题 title_result = soup1.find( 'div' , attrs = { 'class' : 'doc-title' }) doc_title = title_result.get_text() # 得到文档标题 # 拖动滚动条 init_page = driver.find_element_by_xpath( "//div[@class='foldpagewg-text-con']" ) print ( type (init_page), init_page) driver.execute_script( 'arguments[0].scrollIntoView();' , init_page) init_page.click() init = False else : try : page = driver.find_element_by_xpath( "//div[@class='pagerwg-schedule']" ) #print(type(next_page), next_page) next_page = driver.find_element_by_class_name( "pagerwg-button" ) station = driver.find_element_by_xpath( "//div[@class='bottombarwg-root border-none']" ) driver.execute_script( 'arguments[0].scrollIntoView(false);' , station) #js.executeScript("arguments[0].click();",next_page); #sleep(5) '''js = "window.scrollTo(508,600)" driver.execute_script(js)''' next_page.click() except : #结束条件 print ( "找不到元素" ) stop_condition = True #next_page.send_keys(Keys.ENTER) # 遍历所有的txt标签标定的文档,将其空格删除,然后进行保存 content_result = soup1.find_all( 'p' , attrs = { 'class' : 'txt' }) for each in content_result: each_text = each.get_text() if ' ' in each_text: text = each_text.replace( ' ' , '') else : text = each_text # print(each_text) doc_content_list.append(text) # 得到正文内容 sleep( 2 ) # 防止页面加载过慢 if stop_condition is False : doc_title, doc_content_list = find_doc(driver, init) return doc_title, doc_content_list def save(doc_title, doc_content_list): document = Document() heading = document.add_heading(doc_title, 0 ) heading.alignment = WD_ALIGN_PARAGRAPH.CENTER # 居中显示 for each in doc_content_list: document.add_paragraph(each) # 处理字符编码问题 t_title = doc_title.split()[ 0 ] #print(t_title) #document.save('2.docx') document.save( '百度文库-%s.docx' % t_title) print ( "\n\nCompleted: %s.docx, to read." % t_title) driver.quit() if __name__ = = '__main__' : options = webdriver.ChromeOptions() options.add_argument( 'user-agent="Mozilla/5.0 (Linux; Android 4.0.4; \ Galaxy Nexus Build/IMM76B) AppleWebKit/535.19 (KHTML, like Gecko) \ Chrome/18.0.1025.133 Mobile Safari/535.19"' ) #driver = webdriver.Chrome(BROWSER_PATH, chrome_options=options) driver = webdriver.Chrome(chrome_options = options) driver.get(DEST_URL) #JavascriptExecutor js = (JavascriptExecutor) driver; print ( "**********START**********" ) title, content = find_doc(driver, True ) save(title, content) driver.quit() |
希望本文所述对大家Python程序设计有所帮助。