首页 > 解决方案 > 仅查找 BeautifulSoup 元素内的文本

问题描述

我运行这个 python BS 代码:

soup=BeautifulSoup(wd.page_source, 'lxml')
price_divs = soup.find_all("div", class_="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price")
print(price_divs)

这个输出:

<div class="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price">
  <span class="gws-flights-results__carry-on-definitely-not-included gws-flights-results__marker" jsaction="LoTHjf;mouseenter:LoTHjf;mouseleave:QsRKXb" role="button" tabindex="-1"></span> €105</div>
<div class="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price">
  <span class="gws-flights-results__carry-on-definitely-not-included gws-flights-results__marker" jsaction="LoTHjf;mouseenter:LoTHjf;mouseleave:QsRKXb" role="button" tabindex="-1"></span> €105</div>
<div class="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price">€107</div>
<div class="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price">	€107</div>

我希望这只是给我一个所有价格的数组,例如:

[105,107]

谢谢

标签: pythonseleniumweb-scrapingbeautifulsoup

解决方案


如果没有文件样本,请尝试:

soup=BeautifulSoup(wd.page_source, 'lxml')
price_divs = soup.find_all("div", class_="flt-subhead1 gws-flights-results__price gws-flights-results__cheapest-price")

for price in price_divs:
    print(price.text)

为什么:

遍历 div 以便仅查找每个人的文本。


推荐阅读