土地出让.py 8.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207
  1. # -*- coding: utf-8 -*-
  2. """
  3. Created on 2025-04-23
  4. ---------
  5. @summary: 上海市公共资源交易中心
  6. ---------
  7. @author: lzz
  8. """
  9. import feapder
  10. from items.spider_item import DataBakItem
  11. from collections import namedtuple
  12. from feapder.network.selector import Selector
  13. from untils.tools import extract_file_type,get_proxy
  14. import re, time, random
  15. import requests
  16. from untils.attachment import AttachmentDownloader
  17. from feapder.utils.webdriver import WebDriver
  18. def get_html(url):
  19. _kwargs = {}
  20. _kwargs.setdefault("load_images", False)
  21. _kwargs.setdefault("headless", False)
  22. _kwargs.setdefault("driver_type", "CHROME")
  23. _kwargs.setdefault("render_time", 3)
  24. _kwargs.setdefault("usages_local_driver", False)
  25. with WebDriver(**_kwargs) as browser:
  26. try:
  27. browser.get(url)
  28. time.sleep(5)
  29. iframe_html = Selector(browser.page_source).xpath('//div[@id="app"]').extract_first("")
  30. return iframe_html
  31. except Exception as e:
  32. print("iframe获取失败")
  33. return ""
  34. def get_iframe(xmid,proxies=None):
  35. headers = {
  36. "Accept": "application/json, text/plain, */*",
  37. "Accept-Language": "zh-CN,zh;q=0.9,en;q=0.8",
  38. "Cache-Control": "no-cache",
  39. "Connection": "keep-alive",
  40. "Pragma": "no-cache",
  41. "Referer": "https://www.suaee.com/suaeeHome/",
  42. "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36",
  43. "projectType": "suaeeHome",
  44. "sourcecode": "SUAEE"
  45. }
  46. url = "https://www.suaee.com/manageproject/foreign/projectPreview/detail"
  47. params = {
  48. "xmid": xmid
  49. }
  50. try:
  51. response = requests.get(url, headers=headers, params=params, timeout=30, proxies=proxies, verify=False)
  52. file_list = response.json().get('data').get('xgfj')
  53. return file_list
  54. except:
  55. return []
  56. class Shsggzyjyzx(feapder.BiddingListSpider):
  57. def start_callback(self):
  58. Menu = namedtuple('Menu', ['channel', 'code', 'typeone', 'crawl_page'])
  59. self.site = "上海市公共资源交易中心"
  60. self.menus = [
  61. Menu('土地出让', 'sh_shsggzyjyzx_tdcr', 'jyxxtd', 1),
  62. ]
  63. self.headers = {
  64. "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7",
  65. "Accept-Language": "zh-CN,zh;q=0.9,en;q=0.8",
  66. "Cache-Control": "no-cache",
  67. "Connection": "keep-alive",
  68. "Pragma": "no-cache",
  69. "Referer": "https://www.shggzy.com",
  70. "Upgrade-Insecure-Requests": "1",
  71. "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36",
  72. }
  73. self.count = 0
  74. self.proxies = get_proxy()
  75. def start_requests(self):
  76. for menu in self.menus:
  77. start_url = f'https://www.shggzy.com/{menu.typeone}.jhtml'
  78. yield feapder.Request(url=start_url, item=menu._asdict(), use_session=True,
  79. random_user_agent=False, page=1, proxies=False)
  80. def download_midware(self, request):
  81. menu = request.item
  82. typeone = menu.get('typeone')
  83. if request.session == None:
  84. request.session = requests.session()
  85. url = f"https://www.shggzy.com/{typeone}"
  86. res = request.session.get(url, headers=self.headers, proxies = self.proxies, timeout=30, verify=False)
  87. params = {"cExt": f"{res.headers.get('cExt')}"}
  88. request.params = params
  89. request.headers = self.headers
  90. request.proxies = self.proxies
  91. if request.page > 1:
  92. urll = f'https://www.shggzy.com/{typeone}_{request.page}.jhtml'
  93. request.url = urll
  94. def exception_request(self, request, response):
  95. self.proxies = get_proxy()
  96. yield request
  97. def parse(self, request, response):
  98. if self.count > 5:
  99. return
  100. if type(response.text) == bytes:
  101. self.count += 1
  102. yield request
  103. else:
  104. self.count = 0
  105. menu = request.item
  106. info_list = response.xpath('//div[@class="gui-title-bottom"]/ul/li')
  107. for info in info_list:
  108. href_param = info.xpath('./@onclick').extract_first().strip()
  109. href = "https://www.shggzy.com" + "".join(re.findall("window.open\('(.*?)'", href_param, re.S)).strip()
  110. project_code = info.xpath('./span[last()-1]/text()').extract_first("").strip()
  111. dedup_params = href.split('?')[0] + project_code
  112. title = info.xpath('./span[@class="cs-span2"]/text()').extract_first().strip()
  113. create_time = info.xpath('./span[last()]/text()').extract_first().strip()
  114. area = "上海" # 省份
  115. city = "上海市" # 城市
  116. data_item = DataBakItem() # 存储数据的管道
  117. data_item.href = href # 标书链接
  118. data_item.unique_key = (dedup_params,)
  119. data_item.channel = menu.get("channel") # 最上方定义的抓取栏目 (编辑器定的)
  120. data_item.spidercode = menu.get("code") # 最上方定义的爬虫code(编辑器定的)
  121. data_item.title = title # 标题
  122. data_item.publishtime = create_time # 标书发布时间
  123. data_item.site = self.site
  124. data_item.area = area # 城市默认:全国
  125. data_item.city = city # 城市 默认为空
  126. if menu.get('code') in ['sh_shsggzyjyzx_nyys', 'sh_shsggzyjyzx_sfpm']:
  127. data_item.infoformat = 3
  128. time.sleep(random.randint(3, 5))
  129. res = requests.get(href, headers=self.headers, proxies=self.proxies, verify=False, timeout=30)
  130. if res.text:
  131. attachments = {}
  132. iframe_html = ""
  133. root = Selector(res.text)
  134. html = root.xpath('//div[@class="table_1"]').extract_first() # 标书详细内容
  135. if not html:
  136. html = root.xpath('//div[@class="content"]').extract_first()
  137. iframe_url = root.xpath('//iframe[@id="frame-content"]/@src').extract_first("")
  138. if iframe_url:
  139. iframe_html = get_html(iframe_url)
  140. xmid = "".join(re.findall('xmid=(.*?)&',iframe_url))
  141. file_list = get_iframe(xmid=xmid,proxies=self.proxies)
  142. if file_list:
  143. for ff in file_list:
  144. f_name = ff.get('fileName')
  145. fileType = ff.get('fileType')
  146. filePath = ff.get('filePath')
  147. f_url = f"https://www.suaee.com/manageserver/fileDow?type={fileType}&filePath={filePath.replace('/','%2F')}&fileName={f_name}"
  148. f_type = extract_file_type(f_name, f_url)
  149. if f_type:
  150. attachment = AttachmentDownloader().fetch_attachment(
  151. file_name=f_name, file_type=f_type, download_url=f_url)
  152. attachments[str(len(attachments) + 1)] = attachment
  153. files = root.xpath('//div[@class="content-box"]//div[@class="content"]//a[@href]')
  154. if len(files) > 0:
  155. for index, info in enumerate(files):
  156. file_url = info.xpath('./@href').extract_first()
  157. file_name = info.xpath('./text()').extract_first()
  158. file_type = extract_file_type(file_name, file_url)
  159. if file_type and 'file' in file_url:
  160. attachment = AttachmentDownloader().fetch_attachment(
  161. file_name=file_name, file_type=file_type, download_url=file_url)
  162. attachments[str(len(attachments) + 1)] = attachment
  163. if attachments:
  164. data_item.projectinfo = {"attachments": attachments}
  165. extra_html = root.xpath('//div[@id="sectionListDiv"]').extract_first()
  166. if extra_html and html:
  167. html = html.replace(extra_html, '')
  168. data_item.contenthtml = html + iframe_html
  169. yield data_item
  170. time.sleep(random.randint(3, 5))
  171. # 无限翻页
  172. request = self.infinite_pages(request, response)
  173. yield request
  174. if __name__ == "__main__":
  175. Shsggzyjyzx(redis_key="lzz:shsggzyjyzx_tdcr").start()