爬虫基础Scrapy框架之Spalsh渲染
基于spalsh渲染后HTML,通过配置文件解析,入库。 提高了效率,一天可以写几十个配置dict,即完成几十个网站爬虫的编写。
{ "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-规范性文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', # 提取文章列表xpath对象 "title_xpath": "./a", # 提取标题 "title_parse": "./@title", # 提取标题 "title_link_xpath": "./a/@href", # 提取标题链接 "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", # 提取日期时间 "date_parse": "./text()", # 提取日期时间 "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", # link前缀 "config": "{'use_selenium':'False'}" # 其他配置:是否使用selenium(默认使用spalsh)},
# -*- coding: utf-8 -*-'''需求列表:使用任何资讯网站的抓取央行:http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.htmlhttp://www.pbc.gov.cn/tiaofasi/144941/144959/index.html公安部:https://www.mps.gov.cn/n2254314/n2254487/https://www.mps.gov.cn/n2253534/n2253535/index.htmlhttp://www.qth.gov.cn/xxsbxt/sxdw/gajxx/'''from risk_control_info.items import BIgFinanceNewsimport dateparserfrom w3lib.url import canonicalize_urlfrom urllib.parse import urljoinimport scrapyfrom scrapy_splash import SplashRequestfrom risk_control_info.utils import make_md5, generate_text, clean_stringimport rescript = """function main(splash, args) splash.images_enabled = false splash:set_user_agent("{ua}") assert(splash:go(args.url)) assert(splash:wait(args.wait)) return splash:html()end""".format( ua="Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_4) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/74.0.3729.157 Safari/537.36")class BigFinanceAllGovSpider(scrapy.Spider): name = 'big_finance_all_gov' custom_settings = { 'RANDOMIZE_DOWNLOAD_DELAY': True, 'DOWNLOAD_DELAY': 60 / 360.0, 'CONCURRENT_REQUESTS_PER_IP': 8, 'DOWNLOADER_MIDDLEWARES': { 'scrapy_splash.SplashCookiesMiddleware': 723, 'scrapy_splash.SplashMiddleware': 725, 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware': 810, # 'risk_control_info.middlewares.SplashProxyMiddleware': 843, # 代理ip,此方法没成功 'risk_control_info.middlewares.RandomUserAgentMiddleware': 843, 'risk_control_info.middlewares.SeleniumMiddleware': 844 }, # 入库 'ITEM_PIPELINES': { 'risk_control_info.pipelines.RiskControlInfoPipeline': 401, 'risk_control_info.pipelines.MysqlPipeline': 402, }, 'SPIDER_MIDDLEWARES': { 'risk_control_info.middlewares.RiskControlInfoSpiderMiddleware': 543, 'scrapy_splash.SplashDeduplicateArgsMiddleware': 100, }, } def __init__(self, **kwargs): super().__init__() self.env = kwargs.get('env', 'online') def start_requests(self): for target in self.target_info(): if target.get('title_xpath') and target.get('title_link_xpath') \ and target.get('date_xpath') and target.get('article_rows_xpath'): self.logger.info(f"目标网站可配置爬虫信息:{target}") # 使用selenium if target.get("config") and eval(eval(target.get("config")).get('use_selenium')): self.logger.info(f"使用 Selenium 请求 {target['link']}") yield scrapy.Request(url=target['link'], meta={ "target": target, "use_selenium": True }, callback=self.parse, ) else: # 默认使用 Splash self.logger.info(f"使用 Splash 请求 {target['link']}") yield SplashRequest(url=target['link'], meta={"target": target}, callback=self.parse, # endpoint='execute', # args={ # 'lua_source': script, # 'wait': 8}, endpoint='render.json', args={ # 'lua_source': script, # 'proxy': f"http://{proxy_ip_dict['ip']}:{proxy_ip_dict['port']}", 'wait': 10, 'html': 1, 'png': 1 }, ) def parse(self, response): target = response.meta['target'] article_rows = response.xpath(target['article_rows_xpath']) # 遍历所有文字列表 for article_row in article_rows: item = BIgFinanceNews() # 处理标题 _article_row = article_row.xpath(target['title_xpath']) # 定位标题 item['title'] = clean_string( generate_text(_article_row.xpath(target['title_parse']).extract_first().strip())) # 解析标题 # 处理链接 if target.get('prefix'): item['title_link'] = urljoin(target['prefix'], article_row.xpath( target['title_link_xpath']).extract_first()) else: item['title_link'] = article_row.xpath(target['title_link_xpath']).extract_first() # 处理发布日期 # 日期顺序规则 date_order = "YMD" _title_time = article_row.xpath(target['date_xpath']) # 定位:发布时间 _date_str = clean_string( generate_text(_title_time.xpath(target['date_parse']).extract_first())) # 解析:发布时间 if not eval(target.get('date_re_switch')): item['title_time'] = dateparser.parse(_date_str, settings={'DATE_ORDER': date_order}).strftime( "%Y-%m-%d") else: # 使用正则提取时间字符串,存在默认正则表达式 date_re_expression = target.get('date_re_expression', None) _expression = date_re_expression or r"(20\d{2}[-/]?\d{2}[-/]?\d{2})" results = re.findall(r"%s" % _expression, _date_str, re.S) self.logger.info(f"_date_str:{_date_str},results:{results} ") if results: item['title_time'] = dateparser.parse(results[0], settings={'DATE_ORDER': date_order}).strftime( "%Y-%m-%d") else: item['title_time'] = None # 以下写死的 item['bi_channel'] = "gov" item['industry_type'] = f"{target['industry_type']}" item['website_type'] = f"{target['website_type']}" item['url_type'] = f"{target['url_type']}" item['title_hour'] = 0 # 原网站没有发布时间,0代替 item['source_type'] = 0 # 数据来源,0 网站web, 1 微信公众号 item['redis_duplicate_key'] = make_md5(item['title'] + canonicalize_url(item['title_link'])) # 请求详情页 # 使用selenium if target.get("config") and eval(eval(target.get("config")).get('use_selenium')): self.logger.info(f"使用 Selenium 请求 {item['title_link']}") yield scrapy.Request(url=item['title_link'], meta={ "target": target, "use_selenium": True, "item": item }, callback=self.parse_detail, ) else: # 使用 Splash self.logger.info(f"使用 Splash 请求 {item['title_link']}") yield SplashRequest(url=item['title_link'], meta={ "target": target, "item": item }, callback=self.parse_detail, # endpoint='execute', # args={ # 'lua_source': script, # 'wait': 8}, endpoint='render.json', args={ # 'lua_source': script, # 'proxy': f"http://{proxy_ip_dict['ip']}:{proxy_ip_dict['port']}", 'wait': 20, 'html': 1, 'png': 1 }, ) def parse_detail(self, response): self.logger.info(f"处理详情页 {response.url}") item = response.meta['item'] target = response.meta['target'] print(response.xpath(target['content'])) if response.xpath(target['content']): item['content'] = generate_text(response.xpath(target['content']).extract_first()) else: item['content'] = "" yield item @staticmethod def target_info(): ''' 返回目标网站信息 ''' target_list = [ { "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-规范性文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', # 提取文章列表xpath对象 "title_xpath": "./a", # 提取标题 "title_parse": "./@title", # 提取标题 "title_link_xpath": "./a/@href", # 提取标题链接 "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", # 提取日期时间 "date_parse": "./text()", # 提取日期时间 "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", # link前缀 "config": "{'use_selenium':'False'}" # 其他配置:是否使用selenium(默认使用spalsh) }, { "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-其他文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/144959/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', "title_xpath": "./a", "title_parse": "./@title", "title_link_xpath": "./a/@href", "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", "date_parse": "./text()", "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", "config": "{'use_selenium':'False'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "中华人民共和国公安部-规划计划", # 网站模块 "link": "https://www.mps.gov.cn/n2254314/n2254487/", # 访问链接 "article_rows_xpath": '//span/dl/dd', "title_xpath": "./a", "title_parse": "./text()", "title_link_xpath": "./a/@href", "date_re_switch": "True", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./span", "date_parse": "./text()", "content": '//*[@class="arcContent center"]', # 正文HTML xpath "prefix": "https://www.mps.gov.cn/", "config": "{'use_selenium':'True'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "中华人民共和国公安部-公安要闻", # 网站模块 "link": "https://www.mps.gov.cn/n2253534/n2253535/index.html", # 访问链接 "article_rows_xpath": '//span/dl/dd', "title_xpath": "./a", "title_parse": "./text()", "title_link_xpath": "./a/@href", "date_re_switch": "True", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./span", "date_parse": "./text()", "content": '//*[@class="arcContent center"]', # 正文HTML xpath "prefix": "https://www.mps.gov.cn/", "config": "{'use_selenium':'True'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "七台河市人民政府-信息上报系统-市辖单位-公安局", # 网站模块 "link": "http://www.qth.gov.cn/xxsbxt/sxdw/gajxx/", # 访问链接 "article_rows_xpath": '//td[contains(text(), "公安局")]/parent::tr/parent::tbody/parent::table/parent::td/parent::tr/following::tr[1]/td/table//tr/td/a/parent::td/parent::tr', "title_xpath": "./td/a", "title_parse": "./@title", "title_link_xpath": "./td/a/@href", "date_re_switch": "False", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./td[3]", "date_parse": "./text()", "content": '//*[@class="TRS_Editor"]', # 正文HTML xpath "prefix": "http://www.qth.gov.cn/xxsbxt/sxdw/gajxx/", "config": "{'use_selenium':'False'}" }, ] for target in target_list: yield target
IT审计学院
20-06-11 09:23
楼主,想不想参与IT审计项目?
爬虫基础Scrapy框架之Spalsh渲染
xiaotaomi
会员积分:6520
基于spalsh渲染后HTML,通过配置文件解析,入库。 提高了效率,一天可以写几十个配置dict,即完成几十个网站爬虫的编写。
{ "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-规范性文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', # 提取文章列表xpath对象 "title_xpath": "./a", # 提取标题 "title_parse": "./@title", # 提取标题 "title_link_xpath": "./a/@href", # 提取标题链接 "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", # 提取日期时间 "date_parse": "./text()", # 提取日期时间 "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", # link前缀 "config": "{'use_selenium':'False'}" # 其他配置:是否使用selenium(默认使用spalsh)},
# -*- coding: utf-8 -*-'''需求列表:使用任何资讯网站的抓取央行:http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.htmlhttp://www.pbc.gov.cn/tiaofasi/144941/144959/index.html公安部:https://www.mps.gov.cn/n2254314/n2254487/https://www.mps.gov.cn/n2253534/n2253535/index.htmlhttp://www.qth.gov.cn/xxsbxt/sxdw/gajxx/'''from risk_control_info.items import BIgFinanceNewsimport dateparserfrom w3lib.url import canonicalize_urlfrom urllib.parse import urljoinimport scrapyfrom scrapy_splash import SplashRequestfrom risk_control_info.utils import make_md5, generate_text, clean_stringimport rescript = """function main(splash, args) splash.images_enabled = false splash:set_user_agent("{ua}") assert(splash:go(args.url)) assert(splash:wait(args.wait)) return splash:html()end""".format( ua="Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_4) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/74.0.3729.157 Safari/537.36")class BigFinanceAllGovSpider(scrapy.Spider): name = 'big_finance_all_gov' custom_settings = { 'RANDOMIZE_DOWNLOAD_DELAY': True, 'DOWNLOAD_DELAY': 60 / 360.0, 'CONCURRENT_REQUESTS_PER_IP': 8, 'DOWNLOADER_MIDDLEWARES': { 'scrapy_splash.SplashCookiesMiddleware': 723, 'scrapy_splash.SplashMiddleware': 725, 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware': 810, # 'risk_control_info.middlewares.SplashProxyMiddleware': 843, # 代理ip,此方法没成功 'risk_control_info.middlewares.RandomUserAgentMiddleware': 843, 'risk_control_info.middlewares.SeleniumMiddleware': 844 }, # 入库 'ITEM_PIPELINES': { 'risk_control_info.pipelines.RiskControlInfoPipeline': 401, 'risk_control_info.pipelines.MysqlPipeline': 402, }, 'SPIDER_MIDDLEWARES': { 'risk_control_info.middlewares.RiskControlInfoSpiderMiddleware': 543, 'scrapy_splash.SplashDeduplicateArgsMiddleware': 100, }, } def __init__(self, **kwargs): super().__init__() self.env = kwargs.get('env', 'online') def start_requests(self): for target in self.target_info(): if target.get('title_xpath') and target.get('title_link_xpath') \ and target.get('date_xpath') and target.get('article_rows_xpath'): self.logger.info(f"目标网站可配置爬虫信息:{target}") # 使用selenium if target.get("config") and eval(eval(target.get("config")).get('use_selenium')): self.logger.info(f"使用 Selenium 请求 {target['link']}") yield scrapy.Request(url=target['link'], meta={ "target": target, "use_selenium": True }, callback=self.parse, ) else: # 默认使用 Splash self.logger.info(f"使用 Splash 请求 {target['link']}") yield SplashRequest(url=target['link'], meta={"target": target}, callback=self.parse, # endpoint='execute', # args={ # 'lua_source': script, # 'wait': 8}, endpoint='render.json', args={ # 'lua_source': script, # 'proxy': f"http://{proxy_ip_dict['ip']}:{proxy_ip_dict['port']}", 'wait': 10, 'html': 1, 'png': 1 }, ) def parse(self, response): target = response.meta['target'] article_rows = response.xpath(target['article_rows_xpath']) # 遍历所有文字列表 for article_row in article_rows: item = BIgFinanceNews() # 处理标题 _article_row = article_row.xpath(target['title_xpath']) # 定位标题 item['title'] = clean_string( generate_text(_article_row.xpath(target['title_parse']).extract_first().strip())) # 解析标题 # 处理链接 if target.get('prefix'): item['title_link'] = urljoin(target['prefix'], article_row.xpath( target['title_link_xpath']).extract_first()) else: item['title_link'] = article_row.xpath(target['title_link_xpath']).extract_first() # 处理发布日期 # 日期顺序规则 date_order = "YMD" _title_time = article_row.xpath(target['date_xpath']) # 定位:发布时间 _date_str = clean_string( generate_text(_title_time.xpath(target['date_parse']).extract_first())) # 解析:发布时间 if not eval(target.get('date_re_switch')): item['title_time'] = dateparser.parse(_date_str, settings={'DATE_ORDER': date_order}).strftime( "%Y-%m-%d") else: # 使用正则提取时间字符串,存在默认正则表达式 date_re_expression = target.get('date_re_expression', None) _expression = date_re_expression or r"(20\d{2}[-/]?\d{2}[-/]?\d{2})" results = re.findall(r"%s" % _expression, _date_str, re.S) self.logger.info(f"_date_str:{_date_str},results:{results} ") if results: item['title_time'] = dateparser.parse(results[0], settings={'DATE_ORDER': date_order}).strftime( "%Y-%m-%d") else: item['title_time'] = None # 以下写死的 item['bi_channel'] = "gov" item['industry_type'] = f"{target['industry_type']}" item['website_type'] = f"{target['website_type']}" item['url_type'] = f"{target['url_type']}" item['title_hour'] = 0 # 原网站没有发布时间,0代替 item['source_type'] = 0 # 数据来源,0 网站web, 1 微信公众号 item['redis_duplicate_key'] = make_md5(item['title'] + canonicalize_url(item['title_link'])) # 请求详情页 # 使用selenium if target.get("config") and eval(eval(target.get("config")).get('use_selenium')): self.logger.info(f"使用 Selenium 请求 {item['title_link']}") yield scrapy.Request(url=item['title_link'], meta={ "target": target, "use_selenium": True, "item": item }, callback=self.parse_detail, ) else: # 使用 Splash self.logger.info(f"使用 Splash 请求 {item['title_link']}") yield SplashRequest(url=item['title_link'], meta={ "target": target, "item": item }, callback=self.parse_detail, # endpoint='execute', # args={ # 'lua_source': script, # 'wait': 8}, endpoint='render.json', args={ # 'lua_source': script, # 'proxy': f"http://{proxy_ip_dict['ip']}:{proxy_ip_dict['port']}", 'wait': 20, 'html': 1, 'png': 1 }, ) def parse_detail(self, response): self.logger.info(f"处理详情页 {response.url}") item = response.meta['item'] target = response.meta['target'] print(response.xpath(target['content'])) if response.xpath(target['content']): item['content'] = generate_text(response.xpath(target['content']).extract_first()) else: item['content'] = "" yield item @staticmethod def target_info(): ''' 返回目标网站信息 ''' target_list = [ { "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-规范性文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/3581332/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', # 提取文章列表xpath对象 "title_xpath": "./a", # 提取标题 "title_parse": "./@title", # 提取标题 "title_link_xpath": "./a/@href", # 提取标题链接 "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", # 提取日期时间 "date_parse": "./text()", # 提取日期时间 "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", # link前缀 "config": "{'use_selenium':'False'}" # 其他配置:是否使用selenium(默认使用spalsh) }, { "industry_type": "政策", # 行业类别 "website_type": "央行", # 网站/微信公众号名称 "url_type": "中国人民银行-条法司-其他文件", # 网站模块 "link": "http://www.pbc.gov.cn/tiaofasi/144941/144959/index.html", # 访问链接 "article_rows_xpath": '//div[@id="r_con"]//table//tr/td/font[contains(@class, "newslist_style")]', "title_xpath": "./a", "title_parse": "./@title", "title_link_xpath": "./a/@href", "date_re_switch": "False", # 是否使用正则提取日期时间 "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./following-sibling::span[1]", "date_parse": "./text()", "content": '//*[@class="content"]', # 正文HTML xpath "prefix": "http://www.pbc.gov.cn/", "config": "{'use_selenium':'False'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "中华人民共和国公安部-规划计划", # 网站模块 "link": "https://www.mps.gov.cn/n2254314/n2254487/", # 访问链接 "article_rows_xpath": '//span/dl/dd', "title_xpath": "./a", "title_parse": "./text()", "title_link_xpath": "./a/@href", "date_re_switch": "True", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./span", "date_parse": "./text()", "content": '//*[@class="arcContent center"]', # 正文HTML xpath "prefix": "https://www.mps.gov.cn/", "config": "{'use_selenium':'True'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "中华人民共和国公安部-公安要闻", # 网站模块 "link": "https://www.mps.gov.cn/n2253534/n2253535/index.html", # 访问链接 "article_rows_xpath": '//span/dl/dd', "title_xpath": "./a", "title_parse": "./text()", "title_link_xpath": "./a/@href", "date_re_switch": "True", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./span", "date_parse": "./text()", "content": '//*[@class="arcContent center"]', # 正文HTML xpath "prefix": "https://www.mps.gov.cn/", "config": "{'use_selenium':'True'}" }, { "industry_type": "政策", # 行业类别 "website_type": "公安部", # 网站/微信公众号名称 "url_type": "七台河市人民政府-信息上报系统-市辖单位-公安局", # 网站模块 "link": "http://www.qth.gov.cn/xxsbxt/sxdw/gajxx/", # 访问链接 "article_rows_xpath": '//td[contains(text(), "公安局")]/parent::tr/parent::tbody/parent::table/parent::td/parent::tr/following::tr[1]/td/table//tr/td/a/parent::td/parent::tr', "title_xpath": "./td/a", "title_parse": "./@title", "title_link_xpath": "./td/a/@href", "date_re_switch": "False", # 是否使用正则提取日期时间 ( 2020-04-14 ) "date_re_expression": "", # 日期时间正则表达式 "date_xpath": "./td[3]", "date_parse": "./text()", "content": '//*[@class="TRS_Editor"]', # 正文HTML xpath "prefix": "http://www.qth.gov.cn/xxsbxt/sxdw/gajxx/", "config": "{'use_selenium':'False'}" }, ] for target in target_list: yield target
20-06-03 17:41
974
1
回复
楼主,想不想参与IT审计项目?
20-06-11 09:23