如何使用split抓取所有数据?

时间:2020-08-25 08:30:23

标签: python selenium beautifulsoup split web-crawler

像这样爬行数据

abc | abc | ab

ab  | abc | abc

ab  | ab

abc | ab

abc | abc | abc

使用split和if on python

html = driver.page_source
soup = BeautifulSoup(html, 'html.parser'
book_pages = soup.find_all('td', {'class': 'txt lastCol'})

print(book_pages[1].text.split("|")[0])
b_page.append(book_pages[1].text.split("|")[0])

if book_pages[1].text.split("|")[1] != None:
  print(book_pages[1].text.split("|")[1].strip())
  b_weight.append(book_pages[1].text.split("|")[1].strip())

    if book_pages[1].text.split("|")[2] != None:
      print(book_pages[1].text.split("|")[2].strip())
      b_size.append(book_pages[1].text.split("|")[2].strip())

    else:
      continue
else:
 continue

但错误显示 IndexError:列表索引超出范围

如何无误地抓取所有数据?

0 个答案:

没有答案