Training Region-based Object Detectors with Online Hard Example Mining

本文介绍了一种名为OHEM(Online Hard Example Mining)的方法,该方法通过自动选择困难样本简化了训练过程并提高了效率。OHEM消除了传统训练中常用的几个启发式方法和超参数,特别是在基于区域的卷积神经网络(ConvNets)中。实验结果显示,随着训练集规模的增大及难度提升,OHEM能够显著提高平均精度。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

DOI: CVPR 2016

Key Qustion

  • the training set is distinguished by a large imbalance between the number of annotated objects and the number of background examples

Contribution

  • Make training more effective and efficient.
  • OHEM is a simple and intuitive algorithm that eliminates several heuristics and hyperparameters in common use.
  • The candidate examples are subsampled according to a distribution that favors diverse, high loss instances.
  • It removes the need for several heuristics and hyperparameters commonly used in region-based ConvNets.
  • It yields a consistent and significant boosts in mean average precision
  • Its effectiveness increases as the training set becomes larger and more difficult, as demonstrated by results on the MS COCO dataset

Architecture

Fast R-CNN

这里写图片描述

OHEM

这里写图片描述

Experiments

这里写图片描述

这里写图片描述

这里写图片描述

这里写图片描述

Conclusion

  • OHEM eliminates several heuristics and hyperparameters in common use by automatically selecting hard examples, thus simplifying training.
  • Though we used Fast R-CNN throughout this paper, OHEM can be used for training any region-based ConvNet detector.

Unknown Key Words

  • bootstrapping = hard negative mining rely on aforementioned alternation template:(a) for some period of time a fixed model is used to find new examples to add to the active training set; (b) then, for some period of time the model is trained on the fixed active training set.
  • hard positive example = false positive example

Questions

  • However, there is a small caveat: co-located RoIs with high overlap are likely to have correlated losses.
    • use standard non-maximum suppression (NMS) to perform deduplication
  • *

Self-Learning

  • SGD is not suitable for bootstrapping template
  • 2 methods of hard example mining
    • remove easy example and then add some hard example
    • add false positives to dataset to train the model again
  • proposal’s IOU with ground truth is in the interval [bg_lo, 0.5), bg_lo = 0.1 is helpful but ignore some infrequent, but import, difficult background regions.
  • OHEM is robust in case one needs fewer images per batch in order to reduce GPU memory usage.
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值