python爬蟲(chóng)請(qǐng)求庫(kù)httpx和parsel解析庫(kù)的使用測(cè)評(píng)
Python網(wǎng)絡(luò)爬蟲(chóng)領(lǐng)域兩個(gè)最新的比較火的工具莫過(guò)于httpx和parsel了。httpx號(hào)稱下一代的新一代的網(wǎng)絡(luò)請(qǐng)求庫(kù),不僅支持requests庫(kù)的所有操作,還能發(fā)送異步請(qǐng)求,為編寫(xiě)異步爬蟲(chóng)提供了便利。parsel最初集成在著名Python爬蟲(chóng)框架Scrapy中,后獨(dú)立出來(lái)成立一個(gè)單獨(dú)的模塊,支持XPath選擇器, CSS選擇器和正則表達(dá)式等多種解析提取方式, 據(jù)說(shuō)相比于BeautifulSoup,parsel的解析效率更高。
今天我們就以爬取鏈家網(wǎng)上的二手房在售房產(chǎn)信息為例,來(lái)測(cè)評(píng)下httpx和parsel這兩個(gè)庫(kù)。為了節(jié)約時(shí)間,我們以爬取上海市浦東新區(qū)500萬(wàn)元-800萬(wàn)元以上的房產(chǎn)為例。
requests + BeautifulSoup組合
首先上場(chǎng)的是Requests + BeautifulSoup組合,這也是大多數(shù)人剛學(xué)習(xí)Python爬蟲(chóng)時(shí)使用的組合。本例中爬蟲(chóng)的入口url是https://sh.lianjia.com/ershoufang/pudong/a3p5/, 先發(fā)送請(qǐng)求獲取最大頁(yè)數(shù),然后循環(huán)發(fā)送請(qǐng)求解析單個(gè)頁(yè)面提取我們所要的信息(比如小區(qū)名,樓層,朝向,總價(jià),單價(jià)等信息),最后導(dǎo)出csv文件。如果你正在閱讀本文,相信你對(duì)Python爬蟲(chóng)已經(jīng)有了一定了解,所以我們不會(huì)詳細(xì)解釋每一行代碼。
整個(gè)項(xiàng)目代碼如下所示:
# homelink_requests.py # Author: 大江狗 from fake_useragent import UserAgent import requests from bs4 import BeautifulSoup import csv import re import time class HomeLinkSpider(object): def __init__(self): self.ua = UserAgent() self.headers = {"User-Agent": self.ua.random} self.data = list() self.path = "浦東_三房_500_800萬(wàn).csv" self.url = "https://sh.lianjia.com/ershoufang/pudong/a3p5/" def get_max_page(self): response = requests.get(self.url, headers=self.headers) if response.status_code == 200: soup = BeautifulSoup(response.text, 'html.parser') a = soup.select('div[class="page-box house-lst-page-box"]') #使用eval是字符串轉(zhuǎn)化為字典格式 max_page = eval(a[0].attrs["page-data"])["totalPage"] return max_page else: print("請(qǐng)求失敗 status:{}".format(response.status_code)) return None def parse_page(self): max_page = self.get_max_page() for i in range(1, max_page + 1): url = 'https://sh.lianjia.com/ershoufang/pudong/pg{}a3p5/'.format(i) response = requests.get(url, headers=self.headers) soup = BeautifulSoup(response.text, 'html.parser') ul = soup.find_all("ul", class_="sellListContent") li_list = ul[0].select("li") for li in li_list: detail = dict() detail['title'] = li.select('div[class="title"]')[0].get_text() # 2室1廳 | 74.14平米 | 南 | 精裝 | 高樓層(共6層) | 1999年建 | 板樓 house_info = li.select('div[class="houseInfo"]')[0].get_text() house_info_list = house_info.split(" | ") detail['bedroom'] = house_info_list[0] detail['area'] = house_info_list[1] detail['direction'] = house_info_list[2] floor_pattern = re.compile(r'\d{1,2}') # 從字符串任意位置匹配 match1 = re.search(floor_pattern, house_info_list[4]) if match1: detail['floor'] = match1.group() else: detail['floor'] = "未知" # 匹配年份 year_pattern = re.compile(r'\d{4}') match2 = re.search(year_pattern, house_info_list[5]) if match2: detail['year'] = match2.group() else: detail['year'] = "未知" # 文蘭小區(qū) - 塘橋, 提取小區(qū)名和哈快 position_info = li.select('div[class="positionInfo"]')[0].get_text().split(' - ') detail['house'] = position_info[0] detail['location'] = position_info[1] # 650萬(wàn),匹配650 price_pattern = re.compile(r'\d+') total_price = li.select('div[class="totalPrice"]')[0].get_text() detail['total_price'] = re.search(price_pattern, total_price).group() # 單價(jià)64182元/平米, 匹配64182 unit_price = li.select('div[class="unitPrice"]')[0].get_text() detail['unit_price'] = re.search(price_pattern, unit_price).group() self.data.append(detail) def write_csv_file(self): head = ["標(biāo)題", "小區(qū)", "房廳", "面積", "朝向", "樓層", "年份", "位置", "總價(jià)(萬(wàn))", "單價(jià)(元/平方米)"] keys = ["title", "house", "bedroom", "area", "direction", "floor", "year", "location", "total_price", "unit_price"] try: with open(self.path, 'w', newline='', encoding='utf_8_sig') as csv_file: writer = csv.writer(csv_file, dialect='excel') if head is not None: writer.writerow(head) for item in self.data: row_data = [] for k in keys: row_data.append(item[k]) # print(row_data) writer.writerow(row_data) print("Write a CSV file to path %s Successful." % self.path) except Exception as e: print("Fail to write CSV to path: %s, Case: %s" % (self.path, e)) if __name__ == '__main__': start = time.time() home_link_spider = HomeLinkSpider() home_link_spider.parse_page() home_link_spider.write_csv_file() end = time.time() print("耗時(shí):{}秒".format(end-start))
注意:我們使用了fake_useragent, requests和BeautifulSoup,這些都需要通過(guò)pip事先安裝好才能用。
現(xiàn)在我們來(lái)看下爬取結(jié)果,耗時(shí)約18.5秒,總共爬取580條數(shù)據(jù)。
requests + parsel組合
這次我們同樣采用requests獲取目標(biāo)網(wǎng)頁(yè)內(nèi)容,使用parsel庫(kù)(事先需通過(guò)pip安裝)來(lái)解析。Parsel庫(kù)的用法和BeautifulSoup相似,都是先創(chuàng)建實(shí)例,然后使用各種選擇器提取DOM元素和數(shù)據(jù),但語(yǔ)法上稍有不同。Beautiful有自己的語(yǔ)法規(guī)則,而Parsel庫(kù)支持標(biāo)準(zhǔn)的css選擇器和xpath選擇器, 通過(guò)get方法或getall方法獲取文本或?qū)傩灾?,使用起?lái)更方便。
# BeautifulSoup的用法 from bs4 import BeautifulSoup soup = BeautifulSoup(response.text, 'html.parser') ul = soup.find_all("ul", class_="sellListContent")[0] # Parsel的用法, 使用Selector類(lèi) from parsel import Selector selector = Selector(response.text) ul = selector.css('ul.sellListContent')[0] # Parsel獲取文本值或?qū)傩灾蛋咐? selector.css('div.title span::text').get() selector.css('ul li a::attr(href)').get() >>> for li in selector.css('ul > li'): ... print(li.xpath('.//@href').get())
注:老版的parsel庫(kù)使用extract()或extract_first()方法獲取文本或?qū)傩灾?,在新版中已被get()和getall()方法替代。
全部代碼如下所示:
# homelink_parsel.py # Author: 大江狗 from fake_useragent import UserAgent import requests import csv import re import time from parsel import Selector class HomeLinkSpider(object): def __init__(self): self.ua = UserAgent() self.headers = {"User-Agent": self.ua.random} self.data = list() self.path = "浦東_三房_500_800萬(wàn).csv" self.url = "https://sh.lianjia.com/ershoufang/pudong/a3p5/" def get_max_page(self): response = requests.get(self.url, headers=self.headers) if response.status_code == 200: # 創(chuàng)建Selector類(lèi)實(shí)例 selector = Selector(response.text) # 采用css選擇器獲取最大頁(yè)碼div Boxl a = selector.css('div[class="page-box house-lst-page-box"]') # 使用eval將page-data的json字符串轉(zhuǎn)化為字典格式 max_page = eval(a[0].xpath('//@page-data').get())["totalPage"] print("最大頁(yè)碼數(shù):{}".format(max_page)) return max_page else: print("請(qǐng)求失敗 status:{}".format(response.status_code)) return None def parse_page(self): max_page = self.get_max_page() for i in range(1, max_page + 1): url = 'https://sh.lianjia.com/ershoufang/pudong/pg{}a3p5/'.format(i) response = requests.get(url, headers=self.headers) selector = Selector(response.text) ul = selector.css('ul.sellListContent')[0] li_list = ul.css('li') for li in li_list: detail = dict() detail['title'] = li.css('div.title a::text').get() # 2室1廳 | 74.14平米 | 南 | 精裝 | 高樓層(共6層) | 1999年建 | 板樓 house_info = li.css('div.houseInfo::text').get() house_info_list = house_info.split(" | ") detail['bedroom'] = house_info_list[0] detail['area'] = house_info_list[1] detail['direction'] = house_info_list[2] floor_pattern = re.compile(r'\d{1,2}') match1 = re.search(floor_pattern, house_info_list[4]) # 從字符串任意位置匹配 if match1: detail['floor'] = match1.group() else: detail['floor'] = "未知" # 匹配年份 year_pattern = re.compile(r'\d{4}') match2 = re.search(year_pattern, house_info_list[5]) if match2: detail['year'] = match2.group() else: detail['year'] = "未知" # 文蘭小區(qū) - 塘橋 提取小區(qū)名和哈快 position_info = li.css('div.positionInfo a::text').getall() detail['house'] = position_info[0] detail['location'] = position_info[1] # 650萬(wàn),匹配650 price_pattern = re.compile(r'\d+') total_price = li.css('div.totalPrice span::text').get() detail['total_price'] = re.search(price_pattern, total_price).group() # 單價(jià)64182元/平米, 匹配64182 unit_price = li.css('div.unitPrice span::text').get() detail['unit_price'] = re.search(price_pattern, unit_price).group() self.data.append(detail) def write_csv_file(self): head = ["標(biāo)題", "小區(qū)", "房廳", "面積", "朝向", "樓層", "年份", "位置", "總價(jià)(萬(wàn))", "單價(jià)(元/平方米)"] keys = ["title", "house", "bedroom", "area", "direction", "floor", "year", "location", "total_price", "unit_price"] try: with open(self.path, 'w', newline='', encoding='utf_8_sig') as csv_file: writer = csv.writer(csv_file, dialect='excel') if head is not None: writer.writerow(head) for item in self.data: row_data = [] for k in keys: row_data.append(item[k]) # print(row_data) writer.writerow(row_data) print("Write a CSV file to path %s Successful." % self.path) except Exception as e: print("Fail to write CSV to path: %s, Case: %s" % (self.path, e)) if __name__ == '__main__': start = time.time() home_link_spider = HomeLinkSpider() home_link_spider.parse_page() home_link_spider.write_csv_file() end = time.time() print("耗時(shí):{}秒".format(end-start))
現(xiàn)在我們來(lái)看下爬取結(jié)果,爬取580條數(shù)據(jù)耗時(shí)約16.5秒,節(jié)省了2秒時(shí)間。可見(jiàn)parsel比BeautifulSoup解析效率是要高的,爬取任務(wù)少時(shí)差別不大,任務(wù)多的話差別可能會(huì)大些。
httpx同步 + parsel組合
我們現(xiàn)在來(lái)更進(jìn)一步,使用httpx替代requests庫(kù)。httpx發(fā)送同步請(qǐng)求的方式和requests庫(kù)基本一樣,所以我們只需要修改上例中兩行代碼,把requests替換成httpx即可, 其余代碼一模一樣。
from fake_useragent import UserAgent import csv import re import time from parsel import Selector import httpx class HomeLinkSpider(object): def __init__(self): self.ua = UserAgent() self.headers = {"User-Agent": self.ua.random} self.data = list() self.path = "浦東_三房_500_800萬(wàn).csv" self.url = "https://sh.lianjia.com/ershoufang/pudong/a3p5/" def get_max_page(self): # 修改這里把requests換成httpx response = httpx.get(self.url, headers=self.headers) if response.status_code == 200: # 創(chuàng)建Selector類(lèi)實(shí)例 selector = Selector(response.text) # 采用css選擇器獲取最大頁(yè)碼div Boxl a = selector.css('div[class="page-box house-lst-page-box"]') # 使用eval將page-data的json字符串轉(zhuǎn)化為字典格式 max_page = eval(a[0].xpath('//@page-data').get())["totalPage"] print("最大頁(yè)碼數(shù):{}".format(max_page)) return max_page else: print("請(qǐng)求失敗 status:{}".format(response.status_code)) return None def parse_page(self): max_page = self.get_max_page() for i in range(1, max_page + 1): url = 'https://sh.lianjia.com/ershoufang/pudong/pg{}a3p5/'.format(i) # 修改這里把requests換成httpx response = httpx.get(url, headers=self.headers) selector = Selector(response.text) ul = selector.css('ul.sellListContent')[0] li_list = ul.css('li') for li in li_list: detail = dict() detail['title'] = li.css('div.title a::text').get() # 2室1廳 | 74.14平米 | 南 | 精裝 | 高樓層(共6層) | 1999年建 | 板樓 house_info = li.css('div.houseInfo::text').get() house_info_list = house_info.split(" | ") detail['bedroom'] = house_info_list[0] detail['area'] = house_info_list[1] detail['direction'] = house_info_list[2] floor_pattern = re.compile(r'\d{1,2}') match1 = re.search(floor_pattern, house_info_list[4]) # 從字符串任意位置匹配 if match1: detail['floor'] = match1.group() else: detail['floor'] = "未知" # 匹配年份 year_pattern = re.compile(r'\d{4}') match2 = re.search(year_pattern, house_info_list[5]) if match2: detail['year'] = match2.group() else: detail['year'] = "未知" # 文蘭小區(qū) - 塘橋 提取小區(qū)名和哈快 position_info = li.css('div.positionInfo a::text').getall() detail['house'] = position_info[0] detail['location'] = position_info[1] # 650萬(wàn),匹配650 price_pattern = re.compile(r'\d+') total_price = li.css('div.totalPrice span::text').get() detail['total_price'] = re.search(price_pattern, total_price).group() # 單價(jià)64182元/平米, 匹配64182 unit_price = li.css('div.unitPrice span::text').get() detail['unit_price'] = re.search(price_pattern, unit_price).group() self.data.append(detail) def write_csv_file(self): head = ["標(biāo)題", "小區(qū)", "房廳", "面積", "朝向", "樓層", "年份", "位置", "總價(jià)(萬(wàn))", "單價(jià)(元/平方米)"] keys = ["title", "house", "bedroom", "area", "direction", "floor", "year", "location", "total_price", "unit_price"] try: with open(self.path, 'w', newline='', encoding='utf_8_sig') as csv_file: writer = csv.writer(csv_file, dialect='excel') if head is not None: writer.writerow(head) for item in self.data: row_data = [] for k in keys: row_data.append(item[k]) # print(row_data) writer.writerow(row_data) print("Write a CSV file to path %s Successful." % self.path) except Exception as e: print("Fail to write CSV to path: %s, Case: %s" % (self.path, e)) if __name__ == '__main__': start = time.time() home_link_spider = HomeLinkSpider() home_link_spider.parse_page() home_link_spider.write_csv_file() end = time.time() print("耗時(shí):{}秒".format(end-start))
整個(gè)爬取過(guò)程耗時(shí)16.1秒,可見(jiàn)使用httpx發(fā)送同步請(qǐng)求時(shí)效率和requests基本無(wú)差別。
注意:Windows上使用pip安裝httpx可能會(huì)出現(xiàn)報(bào)錯(cuò),要求安裝Visual Studio C++, 這個(gè)下載安裝好就沒(méi)事了。
接下來(lái),我們就要開(kāi)始王炸了,使用httpx和asyncio編寫(xiě)一個(gè)異步爬蟲(chóng)看看從鏈家網(wǎng)上爬取580條數(shù)據(jù)到底需要多長(zhǎng)時(shí)間。
httpx異步+ parsel組合
Httpx厲害的地方就是能發(fā)送異步請(qǐng)求。整個(gè)異步爬蟲(chóng)實(shí)現(xiàn)原理時(shí),先發(fā)送同步請(qǐng)求獲取最大頁(yè)碼,把每個(gè)單頁(yè)的爬取和數(shù)據(jù)解析變?yōu)橐粋€(gè)asyncio協(xié)程任務(wù)(使用async定義),最后使用loop執(zhí)行。
大部分代碼與同步爬蟲(chóng)相同,主要變動(dòng)地方有兩個(gè):
# 異步 - 使用協(xié)程函數(shù)解析單頁(yè)面,需傳入單頁(yè)面url地址 async def parse_single_page(self, url): # 使用httpx發(fā)送異步請(qǐng)求獲取單頁(yè)數(shù)據(jù) async with httpx.AsyncClient() as client: response = await client.get(url, headers=self.headers) selector = Selector(response.text) # 其余地方一樣 def parse_page(self): max_page = self.get_max_page() loop = asyncio.get_event_loop() # Python 3.6之前用ayncio.ensure_future或loop.create_task方法創(chuàng)建單個(gè)協(xié)程任務(wù) # Python 3.7以后可以用戶asyncio.create_task方法創(chuàng)建單個(gè)協(xié)程任務(wù) tasks = [] for i in range(1, max_page + 1): url = 'https://sh.lianjia.com/ershoufang/pudong/pg{}a3p5/'.format(i) tasks.append(self.parse_single_page(url)) # 還可以使用asyncio.gather(*tasks)命令將多個(gè)協(xié)程任務(wù)加入到事件循環(huán) loop.run_until_complete(asyncio.wait(tasks)) loop.close()
整個(gè)項(xiàng)目代碼如下所示:
from fake_useragent import UserAgent import csv import re import time from parsel import Selector import httpx import asyncio class HomeLinkSpider(object): def __init__(self): self.ua = UserAgent() self.headers = {"User-Agent": self.ua.random} self.data = list() self.path = "浦東_三房_500_800萬(wàn).csv" self.url = "https://sh.lianjia.com/ershoufang/pudong/a3p5/" def get_max_page(self): response = httpx.get(self.url, headers=self.headers) if response.status_code == 200: # 創(chuàng)建Selector類(lèi)實(shí)例 selector = Selector(response.text) # 采用css選擇器獲取最大頁(yè)碼div Boxl a = selector.css('div[class="page-box house-lst-page-box"]') # 使用eval將page-data的json字符串轉(zhuǎn)化為字典格式 max_page = eval(a[0].xpath('//@page-data').get())["totalPage"] print("最大頁(yè)碼數(shù):{}".format(max_page)) return max_page else: print("請(qǐng)求失敗 status:{}".format(response.status_code)) return None # 異步 - 使用協(xié)程函數(shù)解析單頁(yè)面,需傳入單頁(yè)面url地址 async def parse_single_page(self, url): async with httpx.AsyncClient() as client: response = await client.get(url, headers=self.headers) selector = Selector(response.text) ul = selector.css('ul.sellListContent')[0] li_list = ul.css('li') for li in li_list: detail = dict() detail['title'] = li.css('div.title a::text').get() # 2室1廳 | 74.14平米 | 南 | 精裝 | 高樓層(共6層) | 1999年建 | 板樓 house_info = li.css('div.houseInfo::text').get() house_info_list = house_info.split(" | ") detail['bedroom'] = house_info_list[0] detail['area'] = house_info_list[1] detail['direction'] = house_info_list[2] floor_pattern = re.compile(r'\d{1,2}') match1 = re.search(floor_pattern, house_info_list[4]) # 從字符串任意位置匹配 if match1: detail['floor'] = match1.group() else: detail['floor'] = "未知" # 匹配年份 year_pattern = re.compile(r'\d{4}') match2 = re.search(year_pattern, house_info_list[5]) if match2: detail['year'] = match2.group() else: detail['year'] = "未知" # 文蘭小區(qū) - 塘橋 提取小區(qū)名和哈快 position_info = li.css('div.positionInfo a::text').getall() detail['house'] = position_info[0] detail['location'] = position_info[1] # 650萬(wàn),匹配650 price_pattern = re.compile(r'\d+') total_price = li.css('div.totalPrice span::text').get() detail['total_price'] = re.search(price_pattern, total_price).group() # 單價(jià)64182元/平米, 匹配64182 unit_price = li.css('div.unitPrice span::text').get() detail['unit_price'] = re.search(price_pattern, unit_price).group() self.data.append(detail) def parse_page(self): max_page = self.get_max_page() loop = asyncio.get_event_loop() # Python 3.6之前用ayncio.ensure_future或loop.create_task方法創(chuàng)建單個(gè)協(xié)程任務(wù) # Python 3.7以后可以用戶asyncio.create_task方法創(chuàng)建單個(gè)協(xié)程任務(wù) tasks = [] for i in range(1, max_page + 1): url = 'https://sh.lianjia.com/ershoufang/pudong/pg{}a3p5/'.format(i) tasks.append(self.parse_single_page(url)) # 還可以使用asyncio.gather(*tasks)命令將多個(gè)協(xié)程任務(wù)加入到事件循環(huán) loop.run_until_complete(asyncio.wait(tasks)) loop.close() def write_csv_file(self): head = ["標(biāo)題", "小區(qū)", "房廳", "面積", "朝向", "樓層", "年份", "位置", "總價(jià)(萬(wàn))", "單價(jià)(元/平方米)"] keys = ["title", "house", "bedroom", "area", "direction", "floor", "year", "location", "total_price", "unit_price"] try: with open(self.path, 'w', newline='', encoding='utf_8_sig') as csv_file: writer = csv.writer(csv_file, dialect='excel') if head is not None: writer.writerow(head) for item in self.data: row_data = [] for k in keys: row_data.append(item[k]) writer.writerow(row_data) print("Write a CSV file to path %s Successful." % self.path) except Exception as e: print("Fail to write CSV to path: %s, Case: %s" % (self.path, e)) if __name__ == '__main__': start = time.time() home_link_spider = HomeLinkSpider() home_link_spider.parse_page() home_link_spider.write_csv_file() end = time.time() print("耗時(shí):{}秒".format(end-start))
現(xiàn)在到了見(jiàn)證奇跡的時(shí)刻了。從鏈家網(wǎng)上爬取了580條數(shù)據(jù),使用httpx編寫(xiě)的異步爬蟲(chóng)僅僅花了2.5秒!!
對(duì)比與總結(jié)
爬取同樣的內(nèi)容,采用不同工具組合耗時(shí)是不一樣的。httpx異步+parsel組合毫無(wú)疑問(wèn)是最大的贏家, requests和BeautifulSoup確實(shí)可以功成身退啦。
- requests + BeautifulSoup: 18.5 秒
- requests + parsel: 16.5秒
- httpx 同步 + parsel: 16.1秒
- httpx 異步 + parsel: 2.5秒
對(duì)于Python爬蟲(chóng),你還有喜歡的庫(kù)嗎?
以上就是python爬蟲(chóng)請(qǐng)求庫(kù)httpx和parsel解析庫(kù)的使用測(cè)評(píng)的詳細(xì)內(nèi)容,更多關(guān)于python httpx和parsel的資料請(qǐng)關(guān)注腳本之家其它相關(guān)文章!
相關(guān)文章
Python運(yùn)算符之Inplace運(yùn)算符的使用教程
Inplace運(yùn)算符的行為與普通運(yùn)算符相似,只是在可變目標(biāo)和不可變目標(biāo)的情況下它們以不同的方式起作用。本文將通過(guò)示例帶大家了解Inplace運(yùn)算符的使用,需要的可以參考一下2022-09-09社區(qū)版pycharm創(chuàng)建django項(xiàng)目的方法(pycharm的newproject左側(cè)沒(méi)有項(xiàng)目選項(xiàng))
pycharm的newproject左側(cè)沒(méi)有出現(xiàn)項(xiàng)目選項(xiàng)的情況下,創(chuàng)建Django項(xiàng)目的解決方法./社區(qū)版pycharm創(chuàng)建django項(xiàng)目的方法,本文給大家介紹的非常詳細(xì),對(duì)大家的學(xué)習(xí)或工作具有一定的參考借鑒價(jià)值,需要的朋友參考下吧2020-09-09python開(kāi)發(fā)之函數(shù)定義實(shí)例分析
這篇文章主要介紹了python開(kāi)發(fā)之函數(shù)定義方法,以實(shí)例形式較為詳細(xì)的分析了Python中函數(shù)的定義與使用技巧,需要的朋友可以參考下2015-11-11基于python實(shí)現(xiàn)垂直爬蟲(chóng)系統(tǒng)的方法詳解
這篇文章主要為大家詳細(xì)介紹了python實(shí)現(xiàn)垂直爬蟲(chóng)系統(tǒng)的方法,文中示例代碼介紹的非常詳細(xì),具有一定的參考價(jià)值,感興趣的小伙伴們可以參考一下,希望能夠給你帶來(lái)幫助2022-03-03Python利用memory_profiler實(shí)現(xiàn)內(nèi)存分析
memory_profiler是第三方模塊,用于監(jiān)視進(jìn)程的內(nèi)存消耗以及python程序內(nèi)存消耗的逐行分析。本文將利用memory_profiler實(shí)現(xiàn)內(nèi)存分析,需要的可以參考一下2022-10-10python opencv實(shí)現(xiàn)證件照換底功能
這篇文章主要為大家詳細(xì)介紹了python opencv實(shí)現(xiàn)證件照換底功能,具有一定的參考價(jià)值,感興趣的小伙伴們可以參考一下2019-08-08用Python獲取攝像頭并實(shí)時(shí)控制人臉的實(shí)現(xiàn)示例
這篇文章主要介紹了用Python獲取攝像頭并實(shí)時(shí)控制人臉的實(shí)現(xiàn)示例,文中通過(guò)示例代碼介紹的非常詳細(xì),對(duì)大家的學(xué)習(xí)或者工作具有一定的參考學(xué)習(xí)價(jià)值,需要的朋友們下面隨著小編來(lái)一起學(xué)習(xí)學(xué)習(xí)吧2019-07-07python使用pgzero進(jìn)行游戲開(kāi)發(fā)
今天要和大家分享的pgzero(pygame zero)是在pygame基礎(chǔ)上做了進(jìn)一步的封裝,使得設(shè)計(jì)一款游戲十分的方便,特別適合少兒編程領(lǐng)域的教學(xué), 與scratch相得益彰。2021-06-06