从特定页面抓取某些URL

时间:2020-08-12 10:54:31

标签: python web-scraping beautifulsoup

我正在尝试从网页中抓取所有与一个主题特别相关的URL。

我正在用漂亮的汤来做。

我当前的尝试是

urls = soup.find_all('a', href=True)

但是我不想在页面上有很多多余的URL。

页面为: https://www.basketball-reference.com/players/

例如,我想抓取所有玩家姓名及其参考代码

 <a href="/players/a/allenra02.html">Ray Allen</a>,

将“ Ray Allen / allenra02”添加到列表中。

如何使用漂亮的汤将必需的前缀添加到url搜索中?例如“玩家/”

2 个答案:

答案 0 :(得分:1)

您可以将编译后的正则表达式用作href=中的.find_all()参数。

例如:

import re
import requests
from bs4 import BeautifulSoup


url = 'https://www.basketball-reference.com/players/'
soup = BeautifulSoup(requests.get(url).content, 'html.parser')

r = re.compile(r'/players/.+/(.*?)\.html')
out = []
for a in soup.find('ul', class_="page_index").find_all('a', href=r):
    out.append('{}/{}'.format(a.get_text(strip=True), r.search(a['href']).group(1)))

from pprint import pprint
pprint(out)

打印:

['Kareem Abdul-Jabbar/abdulka01',
 'Ray Allen/allenra02',
 'LaMarcus Aldridge/aldrila01',
 'Paul Arizin/arizipa01',
 'Carmelo Anthony/anthoca01',
 'Tiny Archibald/architi01',
 'Charles Barkley/barklch01',
 'Kobe Bryant/bryanko01',
 'Larry Bird/birdla01',
 'Walt Bellamy/bellawa01',
 'Rick Barry/barryri01',
 'Chauncey Billups/billuch01',
 'Wilt Chamberlain/chambwi01',
 'Vince Carter/cartevi01',
 'Maurice Cheeks/cheekma01',
 'Stephen Curry/curryst01',

...and so on.

答案 1 :(得分:1)

尝试一下

import requests

url = 'https://www.basketball-reference.com/players/'
soup = BeautifulSoup(requests.get(url).text, "html.parser")

ul = soup.find("ul", attrs={'class':"page_index"})

for li in ul.findAll("li"):
    # ignore the first value (index A,B...)
    for player in li.select("a")[1:]:
        print(
            player.text + "/" + player['href'].split("/")[-1].replace(".html", "")
        )

Kareem Abdul-Jabbar/abdulka01
Ray Allen/allenra02
LaMarcus Aldridge/aldrila01
...
...