scrapy爬虫运行一段时间报错pymysql.err.InterfaceError: (0, ‘’)
有两种方法:
- 在插入前判断是否断连,如果断连则重连
- 在adbapi.ConnectionPool中设置参数cp_reconnect=True
class MysqlTwistedPipeline(object):
def __init__(self, dbpool):
self.dbpool = dbpool
@classmethod
def from_settings(cls, settings):
dbpool = adbapi.ConnectionPool("pymysql", host=settings["MYSQL_HOST"], db=settings["MYSQL_DBNAME"],
user=settings["MYSQL_USER"], passwd=settings["MYSQL_PASSWORD"], charset='utf8',
cursorclass=pymysql.cursors.DictCursor, use_unicode=True,cp_reconnect=True)
print('数据库已链接-----------')
return cls(dbpool)
def process_item(self, item, spider):
# 使用twisted将mysql插入变成异步执行
query = self.dbpool.runInteraction(self.do_insert, item)
query.addErrback(self.handle_error, item, spider)
def handle_error(self, failure, item, spider):
# 处理异步插入的异常
print(failure)
def do_insert(self, cursor, item):
tt = cursor._connection._connection
try:
tt.ping(reconnect=True)
except:
self.dbpool.close()
self.dbpool = adbapi.MysqlConnectionPool().dbpool()
# # 执行具体的插入
insert_sql, items = item.get_insert_sql()
cursor.execute(insert_sql, items)