from bs4 import BeautifulSoup
import time
import requests
from lxml import etree
import xpath_utils
a=time.strftime('%Y-%m-%d %H:%M:%S',time.localtime(time.time()))
print(a)
for i in range(2,10):
baseurl="https://www.cmdy555.com/dianying-{}.html"
headers={"user-agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/94.0.4606.61 Safari/537.36",
"cookie":"UM_distinctid=17c11e1e3b7542-00df64a96a1b2-2343360-1fa400-17c11e1e3b83dd; CNZZDATA1279206515=1539779619-1632379190-%7C1632670796; Hm_lvt_ec14401632059ca87c318150de055c37=1632387589,1632673298; 3276_2319_1.85.41.44=1; 4475_2576_1.85.41.44=1; 3276_2460_1.85.41.44=1; 3276_2601_1.85.41.44=1; 3276_2595_1.85.41.44=1; 3276_2402_1.85.41.44=1; Hm_lpvt_ec14401632059ca87c318150de055c37=1632673324; richviews_3276=g45KANHRn%252BUbhFoG7v%252B6zNySFNiQjyFtCbPt18Jr5cPAQ%252Fqrn3AORivwlEBO3Ly0C6p2pZtSxbiI%252BA%252FyXKDdKQecPrJq16nGlpCkahJMJfpZ43Y4MaN5oIl36%252Fc44ZWBCqOaVoLYjWQgmF6j7Jdytx0SNzH%252FioFwk3s97q9YjC8Zhn2FAYNxZcuHW7gEmDwIp4iLaiSmclgXsIMKm%252FqPPIuhXkmbfCsp2bOJcLSDJ51slGBo1f%252BNkpKbDgArel1b5FpbZe0rpQEr5ElXZZw9EkyMGC1JjnsKl6c5Nz1gbjeZxOeQCpFiKX1g9%252Fl0UtUuaC8SJceFQXrUOfVOGAdMbKEB4k%252BkjC1mbHjifQNtNjvBqjT9iHSET2ieFM8dKIOSF%252Fpk8iOIGCiMdcnGgs8vRTU00K10%252BVO2cE9elCE7RdOXOQHxh%252FX157jJ9tnKvKKT3dmeA%252BQt4iK37LVi3qG%252Fm9pSqE2yzS5LIjGPL7EGn4%252FIoKAOd5patRT4Bf7ezNaU0NBdWBicFn7%252FiddYXIbHLGqkiBFyoDacdYAj0KBKmE%252BV51Pq9k%252BC%252Bb%252FL0XipHt23fmTEfRoO%252FQ7S8FJiM7D4tDI55sLk0dtUTijVbivkl5UApO%252BP54ZBW%252FJbl5TVJGf8OXnaG4eBSrlcrLEmL2Q3g3HrERf%252Ffzxg%252BZVL5%252FmuSXA%253D; 3276_2560_1.85.41.44=1; coupletView_4475=KvyKixajsiF7rhkmBcrJKbrzDhVHlQeTxDi0y%252Fr79yJLO5BbkzE1gV8fNT1SmDIcrG9nGUjbQOKydBc5SYMYtd3jGRpSdVzPOZvRxPsfMXAL%252BYCPQYVk8XpUMAvewiApWhFuE8l73ubcIgwJUTTUF4V5p9np2L5%252F1YKILwhG3FiF1yATaNdfVdCkqPcvKwQr71%252FDgJrpmAh9cw0B7xW0tRRFQ%252FwomrlP3he2Qk1QWu81CfLUremGGlS21GuG%252FpJ47N5rAr%252FmKzXyfuwBen8LZ5bPjrduZw6DeOzglH7zPl%252BK7NN9RJ2rmMKXBQN4Cigyv8OZT%252BqPiaDrzNwu16BUXg%253D%253D"}
url=baseurl.format(i)
response=requests.get(url,headers=headers)
response.encoding="utf-8"
response1=BeautifulSoup(response.text,"html.parser")
response2=response1.find("div",attrs={"class","index-area clearfix"}).find_all("a",attrs={"class","link-hover"})
for q in range(1,2):
demo=etree.HTML(response.text)
response3=demo.xpath('/html/body/div[4]/div[3]/ul/li['str(q)']/a/span/p/text()')
print(response3[0:4])
'''
for x in response2:
res=("https://www.cmdy555.com/"+str(x.get('href')))
print(res)
'''