您的位置:首页技术文章
文章详情页

python实现百度文库自动化爬取

【字号: 日期:2022-06-21 11:54:24浏览:2作者:猪猪
项目介绍

可以下载doc,ppt,pdf.对于doc文档可以下载,doc中的表格无法下载,图片格式的文档也可以下载.ppt和pdf是先下载图片再放到ppt中.只要是可以预览的都可以下载。

已有功能 将可以预览的word文档下载为word文档,如果文档是扫描件,同样支持. 将可以预览的ppt和pdf下载为不可编辑的ppt,因为网页上只有图片,所以理论上无法下载可编辑的版本.环境安装

pip install requestspip install my_fake_useragentpip install python-docxpip install opencv-pythonpip install python-pptxpip install seleniumpip install scrapy

本项目使用的是chromedriver控制chrome浏览器进行数据爬取的的,chromedriver的版本和chrome需要匹配

Windows用看这里

1. 如果你的chrome浏览器版本恰好是87.0.4280,那么恭喜你,你可以直接看使用方式了,因为我下载的chromedriver也是这个版本

2. 如果不是,你需要查看自己的chrome浏览器版本,然后到chromedriver下载地址:http://npm.taobao.org/mirrors/chromedriver/ 这个地址下载对应版本的chromedriver,比如你的浏览器版本是87.0.4280,你就可以找到87.0.4280.20/这个链接,如果你是windows版本然后选择chromedriver_win32.zip进行下载解压。千万不要下载LASEST——RELEASE87.0.4280这个链接,这个链接没有用,之前有小伙伴走过弯路的,注意一下哈。

3. 用解压好的chromedriver.exe替换原有文件,然后跳到使用方式

ubuntu用户看这里

讲道理,你已经用ubuntu了,那位就默认你是大神,你只要根据chrome的版本下载对应的chromdriver(linux系统的),然后把chromedriver的路径改称你下载解压的文件路径就好了,然后跳到使用方式。哈哈哈,我这里就偷懒不讲武德啦

使用方式:

把代码中的url改为你想要下载的链接地址,脚本会自动文档判断类型,并把在当前目录新建文件夹并把文件下载到当前目录。

主要代码

import osimport timefrom selenium import webdriverfrom selenium.webdriver.common.desired_capabilities import DesiredCapabilitiesfrom scrapy import Selectorimport requestsfrom my_fake_useragent import UserAgentimport docxfrom docx.shared import Inchesimport cv2from pptx import Presentationfrom pptx.util import Inches#dows是的chromedriverchromedriver_path = './chromedriver.exe'#用ubuntu的chromedriver# chromedriver_path = './chromedriver'doc_dir_path = './doc'ppt_dir_path = './ppt'# url = 'https://wenku.baidu.com/view/4410199cb0717fd5370cdc2e.html?fr=search'# doc_txt p# url = 'https://wenku.baidu.com/view/4d18916f7c21af45b307e87101f69e314332fa36.html' # doc_txt span# url = 'https://wenku.baidu.com/view/dea519c7e53a580216fcfefa.html?fr=search' # doc_txt span br# url = ’https://wk.baidu.com/view/062edabeb6360b4c2e3f5727a5e9856a5712262d?pcf=2&bfetype=new’ # doc_img# url = 'https://wenku.baidu.com/view/2af6de34a7e9856a561252d380eb6294dd88228d'# vip限定doc# url = 'https://wenku.baidu.com/view/3de365cc6aec0975f46527d3240c844769eaa0aa.html?fr=search' #ppt# url = 'https://wenku.baidu.com/view/18a8bc08094e767f5acfa1c7aa00b52acec79c55'#pdf# url = 'https://wenku.baidu.com/view/bbe27bf21b5f312b3169a45177232f60dccce772'# url = 'https://wenku.baidu.com/view/5cb11d096e1aff00bed5b9f3f90f76c660374c24.html?fr=search'# url = 'https://wenku.baidu.com/view/71f9818fef06eff9aef8941ea76e58fafab045a6.html'# url = 'https://wenku.baidu.com/view/ffc6b32a68eae009581b6bd97f1922791788be69.html'url = 'https://wenku.baidu.com/view/d4d2e1e3122de2bd960590c69ec3d5bbfd0adaa6.html'class DownloadImg(): def __init__(self):self.ua = UserAgent() def download_one_img(self, img_url, saved_path):# 下载图片header = { 'User-Agent': '{}'.format(self.ua.random().strip()), ’Connection’: ’close’}r = requests.get(img_url, headers=header, stream=True)print('请求图片状态码 {}'.format(r.status_code)) # 返回状态码if r.status_code == 200: # 写入图片 with open(saved_path, mode='wb') as f:f.write(r.content) print('download {} success!'.format(saved_path))del rreturn saved_pathclass StartChrome(): def __init__(self):mobile_emulation = {'deviceName': 'Galaxy S5'}capabilities = DesiredCapabilities.CHROMEcapabilities[’loggingPrefs’] = {’browser’: ’ALL’}options = webdriver.ChromeOptions()options.add_experimental_option('mobileEmulation', mobile_emulation)self.brower = webdriver.Chrome(executable_path=chromedriver_path, desired_capabilities=capabilities, chrome_options=options)# 启动浏览器,打开需要下载的网页self.brower.get(url)self.download_img = DownloadImg() def click_ele(self, click_xpath):# 单击指定控件click_ele = self.brower.find_elements_by_xpath(click_xpath)if click_ele: click_ele[0].location_once_scrolled_into_view # 滚动到控件位置 self.brower.execute_script(’arguments[0].click()’, click_ele[0]) # 单击控件,即使控件被遮挡,同样可以单击 def judge_doc(self, contents):# 判断文档类别p_list = ’’.join(contents.xpath('./text()').extract())span_list = ’’.join(contents.xpath('./span/text()').extract())# # if span_list# if len(span_list)>len(p_list):# xpath_content_one = './br/text()|./span/text()|./text()'# elif len(span_list)<len(p_list):# # xpath_content_one = './br/text()|./text()'# xpath_content_one = './br/text()|./span/text()|./text()'if len(span_list)!=len(p_list): xpath_content_one = './br/text()|./span/text()|./text()'else: xpath_content_one = './span/img/@src'return xpath_content_one def create_ppt_doc(self, ppt_dir_path, doc_dir_path):# 点击关闭开通会员按钮xpath_close_button = '//div[@class=’na-dialog-wrap show’]/div/div/div[@class=’btn-close’]'self.click_ele(xpath_close_button)# 点击继续阅读xpath_continue_read_button = '//div[@class=’foldpagewg-icon’]'self.click_ele(xpath_continue_read_button)# 点击取消打开百度app按钮xpath_next_content_button = '//div[@class=’btn-wrap’]/div[@class=’btn-cancel’]'self.click_ele(xpath_next_content_button)# 循环点击加载更多按钮,直到显示全文click_count = 0while True: # 如果到了最后一页就跳出循环 if self.brower.find_elements_by_xpath('//div[@class=’pagerwg-loadSucc hide’]') or self.brower.find_elements_by_xpath('//div[@class=’pagerwg-button’ and @style=’display: none;’]'):break # 点击加载更多 xpath_loading_more_button = '//span[@class=’pagerwg-arrow-lower’]' self.click_ele(xpath_loading_more_button) click_count += 1 print('第{}次点击加载更多!'.format(click_count)) # 等待一秒,等浏览器加载 time.sleep(1.5)# 获取html内容sel = Selector(text=self.brower.page_source)#判断文档类型xpath_content = '//div[@class=’content singlePage wk-container’]/div/p/img/@data-loading-src|//div[@class=’content singlePage wk-container’]/div/p/img/@data-src'contents = sel.xpath(xpath_content).extract()if contents:#如果是ppt self.create_ppt(ppt_dir_path, sel)else:#如果是doc self.create_doc(doc_dir_path, sel)# a = 3333# return sel def create_ppt(self, ppt_dir_path, sel):# 如果文件夹不存在就创建一个if not os.path.exists(ppt_dir_path): os.makedirs(ppt_dir_path)SLD_LAYOUT_TITLE_AND_CONTENT = 6 # 6代表ppt模版为空prs = Presentation() # 实例化ppt# # 获取完整html# sel = self.get_html_data()# 获取标题xpath_title = '//div[@class=’doc-title’]/text()'title = ''.join(sel.xpath(xpath_title).extract()).strip()# 获取内容xpath_content_p = '//div[@class=’content singlePage wk-container’]/div/p/img'xpath_content_p_list = sel.xpath(xpath_content_p)xpath_content_p_url_list=[]for imgs in xpath_content_p_list: xpath_content = './@data-loading-src|./@data-src|./@src' contents_list = imgs.xpath(xpath_content).extract() xpath_content_p_url_list.append(contents_list)img_path_list = [] # 保存下载的图片路径,方便后续图片插入ppt和删除图片# 下载图片到指定目录for index, content_img_p in enumerate(xpath_content_p_url_list): p_img_path_list=[] for index_1,img_one in enumerate(content_img_p):one_img_saved_path = os.path.join(ppt_dir_path, '{}_{}.jpg'.format(index,index_1))self.download_img.download_one_img(img_one, one_img_saved_path)p_img_path_list.append(one_img_saved_path) p_img_max_shape = 0 for index,p_img_path in enumerate(p_img_path_list):img_shape = cv2.imread(p_img_path).shapeif p_img_max_shape<img_shape[0]: p_img_max_shape = img_shape[0] index_max_img = index img_path_list.append(p_img_path_list[index_max_img])print(img_path_list)# 获取下载的图片中最大的图片的尺寸img_shape_max=[0,0]for img_path_one in img_path_list: img_path_one_shape = cv2.imread(img_path_one).shape if img_path_one_shape[0]>img_shape_max[0]:img_shape_max = img_path_one_shape# 把图片统一缩放最大的尺寸for img_path_one in img_path_list: cv2.imwrite(img_path_one,cv2.resize(cv2.imread(img_path_one),(img_shape_max[1],img_shape_max[0])))# img_shape_path = img_path_list[0]# 获得图片的尺寸# img_shape = cv2.imread(img_shape_path).shape# 把像素转换为ppt中的长度单位emu,默认dpi是720# 1厘米=28.346像素=360000# 1像素 = 12700emuprs.slide_width = img_shape_max[1] * 12700 # 换算单位prs.slide_height = img_shape_max[0] * 12700for img_path_one in img_path_list: left = Inches(0) right = Inches(0) # width = Inches(1) slide_layout = prs.slide_layouts[SLD_LAYOUT_TITLE_AND_CONTENT] slide = prs.slides.add_slide(slide_layout) pic = slide.shapes.add_picture(img_path_one, left, right, ) print('insert {} into pptx success!'.format(img_path_one)) # os.remove(img_path_one)for root,dirs,files in os.walk(ppt_dir_path): for file in files:if file.endswith('.jpg'): img_path = os.path.join(root,file) os.remove(img_path)prs.save(os.path.join(ppt_dir_path, title + '.pptx'))print('download {} success!'.format(os.path.join(ppt_dir_path, title + '.pptx'))) def create_doc(self, doc_dir_path, sel):# 如果文件夹不存在就创建一个if not os.path.exists(doc_dir_path): os.makedirs(doc_dir_path)# # 获取完整html# sel = self.get_html_data()# 获取标题xpath_title = '//div[@class=’doc-title’]/text()'title = ''.join(sel.xpath(xpath_title).extract()).strip()document = docx.Document() # 创建word文档document.add_heading(title, 0) # 添加标题# 获取文章内容xpath_content = '//div[contains(@data-id,’div_class_’)]//p'# xpath_content = '//div[contains(@data-id,’div_class_’)]/p'contents = sel.xpath(xpath_content)# 判断内容类别xpath_content_one = self.judge_doc(contents)if xpath_content_one.endswith('text()'): # 如果是文字就直接爬 for content_one in contents:one_p_list = content_one.xpath(xpath_content_one).extract()p_txt = ''for p in one_p_list: if p==' ':p_txt += (’n’+p) else:p_txt += p# content_txt_one = ’*’.join(content_one.xpath(xpath_content_one).extract())pp = document.add_paragraph(p_txt) document.save(os.path.join(doc_dir_path, ’{}.docx’.format(title))) print('download {} success!'.format(title))elif xpath_content_one.endswith('@src'): # 如果是图片就下载图片 for index, content_one in enumerate(contents.xpath(xpath_content_one).extract()):# 获取图片下载路径content_img_one_url = ’https:’ + content_one# 保存图片saved_image_path = self.download_img.download_one_img(content_img_one_url, os.path.join(doc_dir_path,'{}.jpg'.format( index)))document.add_picture(saved_image_path, width=Inches(6)) # 在文档中加入图片os.remove(saved_image_path) # 删除下载的图片 document.save(os.path.join(doc_dir_path, ’{}.docx’.format(title))) # 保存文档到指定位置 print('download {} success!'.format(title))if __name__ == '__main__': start_chrome = StartChrome() # start_chrome.create_doc_txt(doc_dir_path) start_chrome.create_ppt_doc(ppt_dir_path, doc_dir_path)项目地址

https://github.com/siyangbing/baiduwenku

以上就是python实现百度文库自动化爬取的详细内容,更多关于python 爬取百度文库的资料请关注好吧啦网其它相关文章!

标签: 百度 Python
相关文章: