启动文件main.py
from scrapy.cmdline import execute
execute('scrapy crawl wangyi'.split())
执行spider文件下的爬取文件
# -*- coding: utf-8 -*-
import scrapy,re
from ..piaot import * #导入自定义包
from ..items import wangye_mysql
class WangyiSpider(scrapy.Spider):
name = 'wangyi'
allowed_domains = ['news.163.com']
# 初始化
def start_requests(self):
# 输入页数
ye=int(input('请输入页数:'))
# 循环页数
for i in range(1,ye+1):
if i == 1:
url='http://temp.163.com/special/00804KVA/cm_guonei.js?callback=data_callback'
else