点击跳转到总目录
本篇只记录程序点击跳转项目
直接上代码
- 此项目主要为了学习RFM模型,R(近度)F(频度)M(额度),我把目标盯上了淘宝,京东这种购物网站,但是水平太菜,于是退而其次打算爬取当当网试试
- 参考写了这个爬虫之后发现获取到的数据并不理想,故放弃
import requests
from lxml import etree
import pandas as pd
test_url = 'http://search.dangdang.com/?key='+ '数据分析'
content_page = requests.get(test_url).text
print(content_page[:1000])
def content(content_page):
books = []
page = etree.HTML(content_page)
book_name = page.xpath('//li/p/a[@name="itemlist-title"]/@title')
pub_info = page.xpath('//li/p[@class="search_book_author"]')
pub_info = [book_pub.xpath('string(.)') for book_pub in pub_info]
price_now = page.xpath('//li//span[@class="search_now_price"]/text()')
stars = page.xpath('//li/p[@class="search_star_line"]/span[@class="search_star_black"]/span/@style')
comment_num = page.xpath('//li/p[@class="search_star_line"]/a[@class="search_comment_num"]/text()')
for book in zip(book_name, pub_info, price_now, stars, comment_num):
books.append(list(book))
return books
test_url = 'http://search.dangdang.com/?key='+ '数据分析'
content_page = requests.get(test_url).text
books = content(content_page)
print(books[:5])
books_df = pd.DataFrame(data=books,columns=["书名","出版信息","当前价格","星级","评论数"])
books_df.to_csv("./2.csv",encoding="utf8",sep="\t",index=None)
books_df.to_excel("./2.xlsx",sheet_name="sheet1",index=False,encoding="utf-8")
