|
本帖最后由 xjl565135022 于 2024-9-12 17:43 编辑
- <div class="blockcode"><blockquote>import time
- import requests
- from bs4 import BeautifulSoup
- from openpyxl import load_workbook,Workbook
- from concurrent.futures import ThreadPoolExecutor
- def data_get(url,index):
- headers = {'user-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/121.0.0.0 Safari/537.36',
- 'cookie': 'qcc_did=da56d80d-a133-40cf-a2e1-bf4664d6f2fd; UM_distinctid=191e406fef2e2d-0dc8b31fb5cea2-17525637-183990-191e406fef31805; QCCSESSID=ceba1c940ef43be4f55775d7ef; tfstk=fdeS_jjNvye4qr1P1HIqGNq-JrHQFy6ZRHiLjkpyvYHJJeEtu0yPrk3QRru0RXjo823Q8kFP0OWaquDnpW_NQORhjlRuPU3JwMnxbc84TNfUquDHm0aTjyqkOe6DryHLJxhxxcHpwvhdcxnn2vp-9Ldvkq0xJvhJpEdxYcLpe4nRco3mvv3KwHKx0Tg7Vi39ZmPwbCESlpp6j2lRvkAeULrrVbwL2qiSV-iSNVh2nNWaClE8IDwhXBgQ-7Urmy61lAFglJGtd99mUlNYPjN5RKG0Ml2KGSb6slqLjWhbdUIzWrFQJmVcRBg_9uPg3j-pMbFgz5hb3a9mWSEbYjNcxQuUsDUnsJ7ME4ELjRVZCTTiVSEshgSDQV_TEBtjspij7isXtBfjJzbv4xi2xbnm01jfcUanwmmjmisXtBc-mmrCciT7-; acw_tc=0a47318a17261321581815187e007a4def7a9ca4cc81449bb13b9f1c99dc6a; CNZZDATA1254842228=1018870652-1726107549-%7C1726132160'}
- response=requests.get(url,headers=headers)
- soup = BeautifulSoup(response.text, 'lxml') # 格式化数据
- data = soup.find_all('span', class_="max-address copy-value address-map") # 获取数据
- print(data[0].text)
- ws.cell(index + 1, 4).value = data[0].text
- data = soup.find_all('span', class_="f tel") # 获取数据
- print(data[0].find("span").text)
- ws.cell(index + 1, 5).value = data[0].find("span").text
- urls=[]
- file_path='/Users/liaogege/Desktop/获取/清单.xlsx'
- wb=load_workbook(file_path)
- ws=wb.active
- index = 0
- for index,cell in enumerate(ws["B"]):
- if index==0:
- continue
- if cell.value:
- urls.append(f'https://www.qcc.com/web/search?key={cell.value}')
- index = 0
- for url in urls:
- if url:
- index+=1
- time.sleep(1) # 防止被封
- with ThreadPoolExecutor(max_workers=len(url)) as executor:
- executor.submit(data_get,url,index)
- wb.save(file_path)
复制代码
|
评分
-
1
查看全部评分
-
|