scrapy爬虫数据入库后有重复数据

版权声明:本文为博主原创文章,未经博主允许不得转载。 https://blog.csdn.net/haoyuexihuai/article/details/83043049
import copy

#10创建数据库 
 def insertProgressReportDrillById(self, item):
     sql = "insert into rpm_project_progress_fromxy(sum_report_progress,sum_production,proj_code) values(%s,%s,%s)"
     #对象拷贝,深拷贝  --- 这里是解决办法
     asynItem = copy.deepcopy(item)
     query = self.dbpool.runInteraction(self._conditional_insertProgressReportDrillById, sql, asynItem)
     query.addErrback(self._handle_error)
     return item

 #10写入数据库中 
 def _conditional_insertProgressReportDrillById(self, tx, sql, item):
     params = (item['sum_report_progress'],item['sum_production'],item['proj_code'])
     tx.execute(sql, params)

猜你喜欢

转载自blog.csdn.net/haoyuexihuai/article/details/83043049