爬虫爬到371页就无法继续了,显示list index out of range
```python
from selenium import webdriver
from selenium.webdriver import ActionChains
from selenium.webdriver.common.by import By
from lxml import etree
from time import sleep
import pandas as pd
from bs4 import BeautifulSoup
def get_row(li):
list1 = []
l1_1 = li.find_all('a', limit=2)
l1_2 = li.find_all('th')
for x in l1_1:
list1.append(x.string)
for y in l1_2:
list1.append(y.string)
return list1
# 实例化一个浏览器对象
bro = webdriver.Chrome(executable_path='./chromedriver')
# 让浏览器发起一个指定url的请求
bro.get('http://vip.stock.finance.sina.com.cn/fund_center/index.html#hbphall')
# 获取浏览器当前界面的动态页面源码数据
page_text = bro.page_source
# 解析企业名称
soup = BeautifulSoup(page_text, "html.parser")
# 建立空列表
list_ji = []
# 开始爬取数据
for i in range(1, 3, 1):
if i <= 414:
j = 0
for j in range(0, 40, 1):
l1 = soup.find_all('tr', class_='red')
list_ji.append(get_row(l1[j]))
# 翻页
next_button = bro.find_element(By.LINK_TEXT, '下一页')
actions = ActionChains(bro)
actions.click(next_button).perform()
# 获取浏览器当前界面的动态页面源码数据
page_text = bro.page_source
# 解析企业名称
soup = BeautifulSoup(page_text, "html.parser")
else:
for k in range(0, 30, 1):
l1 = soup.find_all('tr', class_='red')
list_ji.append(get_row(l1[k]))
column = ["基金代码", "基金名称", "单位净值", "累计净值", "近三个月(%)", "近六个月(%)", "近一年(%)",
"今年以来(%)", "成立以来(%)"]
result = pd.DataFrame(list_ji, columns=column)
pd.set_option('display.max_rows', None) # 显示pandas所有行
# result.columns = ["基金代码", "基金名称", "单位净值", "累计净值", "近三个月(%)", "近六个月(%)", "近一年(%)", "今年以来(%)", "成立以来(%)"]
print(result)
bro.quit()
```