目录
1.前言
2.头文件ORBmartcher.h
/该类负责特征点与特征点之间,地图点与特征点之间通过投影关系、词袋模型或者Sim3位姿匹配
class ORBmatcher
{
public:
/**
* 找到在 以x, y为中心,边长为2r的方形内且尺度在[minLevel, maxLevel]的特征点
* @param nnratio 匹配特征点时,确定时候最好匹配与次好匹配差距的阈值。其值越小,其匹配越精确
* @param checkOri 是否开启匹配点的方向分类
*/
ORBmatcher(float nnratio=0.6, bool checkOri=true);
// Computes the Hamming distance between two ORB descriptors
static int DescriptorDistance(const cv::Mat &a, const cv::Mat &b);
// Search matches between Frame keypoints and projected MapPoints. Returns number of matches
// Used to track the local map (Tracking)
//将F里的特征值与vpMapPoints进行匹配,通过投影加速
//返回通过此函数匹配成功的数量
int SearchByProjection(Frame &F, const std::vector<MapPoint*> &vpMapPoints, const float th=3);
// Project MapPoints tracked in last frame into the current frame and search matches.
// Used to track from previous frame (Tracking)
/**根据上一帧LastFrame的特征点以及所对应的mappoint信息,寻找当前帧的哪些特征点与哪些mappoint的匹配联系
* 根据上一帧特征点对应的3D点投影的位置缩小特征点匹配范围
* @param CurrentFrame 当前帧
* @param LastFrame 上一帧
* @param th 控制特征搜索框的大小阈值
* @param bMono 是否为单目
* @return 成功匹配的数量
*/
int SearchByProjection(Frame &CurrentFrame, const Frame &LastFrame, const float th, const bool bMono);
// Project MapPoints seen in KeyFrame into the Frame and search matches.
// Used in relocalisation (Tracking)
/**在Tracking里的relocalisation中使用
* CurrentFrame中特征点已经匹配好一些mappoint在sAlreadyFound中,通过此函数将pKF悉数投影到CurrentFrame再就近搜索特征点进行匹配
* 也就是说CurrentFrame想通过这个函数在pKF的mappoint集合中匹配上更多的mappoint点
* @param CurrentFrame 当前帧
* @param pKF
* @param sAlreadyFound CurrentFrame已经匹配上的mappoint
* @param th 控制特征搜索框的大小阈值
* @param ORBdist pKF中的mappoint是否能和CurrentFrame匹配成功的描述子距离的阈值
* @return 成功匹配的数量
*/
int SearchByProjection(Frame &CurrentFrame, KeyFrame* pKF, const std::set<MapPoint*> &sAlreadyFound, const float th, const int ORBdist);
// Project MapPoints using a Similarity Transformation and search matches.
// Used in loop detection (Loop Closing)
int SearchByProjection(KeyFrame* pKF, cv::Mat Scw, const std::vector<MapPoint*> &vpPoints, std::vector<MapPoint*> &vpMatched, int th);
// Search matches between MapPoints in a KeyFrame and ORB in a Frame.
// Brute force constrained to ORB that belong to the same vocabulary node (at a certain level)
// Used in Relocalisation and Loop Detection
/**
* 针对pKF中有对应mappoint的那些特征点,和F中的特征点进行匹配
* 利用dbow进行匹配加速
* 通过距离阈值、比例阈值和角度投票进行剔除误匹配
* @param pKF KeyFrame
* @param F Current Frame
* @param vpMapPointMatches 大小为F特征点数量,输出F中MapPoints对应的匹配,NULL表示未匹配
* @return 成功匹配的数量
*/
int SearchByBoW(KeyFrame *pKF, Frame &F, std::vector<MapPoint*> &vpMapPointMatches);
// 匹配pKF1与pKF2之间的特征点并通过bow加速,vpMatches12是pKF1匹配特征点对应的MapPoints
int SearchByBoW(KeyFrame *pKF1, KeyFrame* pKF2, std::vector<MapPoint*> &vpMatches12);
// Matching for the Map Initialization (only used in the monocular case)
/**
* 搜索F1和F2之间的匹配点放在vnMatches12。如果mbCheckOrientation,则将F1中匹配成功的
* 特征点按照其角度分类在rotHist中。
* 计算出rotHist,vnMatches12
* @param F1 参考帧
* @param F2 当前帧
* @param vbPrevMatched F1中待匹配的特征点
* @param vnMatches12 输出F1中特征点匹配情况,大小是F1的特征点数量。其中-1表示未匹配,大于0表示匹配的特征点在F2中的序号
* @param windowSize 加速匹配时用到的方形边长
* @return 成功匹配的数量
*/
int SearchForInitialization(Frame &F1, Frame &F2, std::vector<cv::Point2f> &vbPrevMatched, std::vector<int> &vnMatches12, int windowSize=10);
// Matching to triangulate new MapPoints. Check Epipolar Constraint.
// 匹配pKF1与pKF2之间的未被匹配的特征点并通过bow加速,并校验是否符合对级约束。vMatchedPairs匹配成功的特征点在各自关键帧中的id。
int SearchForTriangulation(KeyFrame *pKF1, KeyFrame* pKF2, cv::Mat F12,
std::vector<pair<size_t, size_t> > &vMatchedPairs, const bool bOnlyStereo);
// Search matches between MapPoints seen in KF1 and KF2 transforming by a Sim3 [s12*R12|t12]
// In the stereo and RGB-D case, s12=1
int SearchBySim3(KeyFrame* pKF1, KeyFrame* pKF2, std::vector<MapPoint *> &vpMatches12, const float &s12, const cv::Mat &R12, const cv::Mat &t12, const float th);
// Project MapPoints into KeyFrame and search for duplicated MapPoints.
/**
* 将vpMapPoints中的mappoint与pKF的特征点进行匹配,若匹配的特征点已有mappoint与其匹配,
* 则选择其一与此特征点匹配,并抹去没有选择的那个mappoint,此为mappoint的融合
* @param radius 为在vpMapPoints投影到pKF搜索待匹配的特征点时的方框边长
* @return 融合的mappoint的数量
*/
int Fuse(KeyFrame* pKF, const vector<MapPoint *> &vpMapPoints, const float th=3.0);
// Project MapPoints into KeyFrame using a given Sim3 and search for duplicated MapPoints.
//vpPoints通过Scw投影到pKF,与pKF中的特征点匹配。如果匹配的pKF中的特征点本身有就的匹配mappoint,就用vpPoints替代它。
//vpReplacePoint大小与vpPoints一致,储存着被替换下来的mappoint
int Fuse(KeyFrame* pKF, cv::Mat Scw, const std::vector<MapPoint*> &vpPoints, float th, vector<MapPoint *> &vpReplacePoint);
public:
//匹配特征点时,描述子距离的阈值。特征点间描述子小于此值才考虑匹配
static const int TH_LOW;
static const int TH_HIGH;
//按照匹配特征点之间的角度分类匹配特征点的数量
static const int HISTO_LENGTH;
protected:
//判断kp1,与kp2在基础矩阵F12下是否复合对极约束
bool CheckDistEpipolarLine(const cv::KeyPoint &kp1, const cv::KeyPoint &kp2, const cv::Mat &F12, const KeyFrame *pKF);
//根据观测角的cos值确定搜索区域的半径
float RadiusByViewingCos(const float &viewCos);
//找出数组histo中,vector.size()数量最大的前三位。也就是角度范围最多的前三位。
void ComputeThreeMaxima(std::vector<int>* histo, const int L, int &ind1, int &ind2, int &ind3);
//匹配特征点时,确定时候最好匹配与次好匹配差距的阈值。其值越小,其匹配越精确
float mfNNratio;
//是否开启匹配点角度差与其他大多数匹配点角度差差异较大的匹配点
bool mbCheckOrientation;
};
3.源文件ORBmartcher
3.1.SearchByProjection()
int ORBmatcher::SearchByProjection(Frame &F, const vector<MapPoint*> &vpMapPoints, const float th)
{
int nmatches=0;
const bool bFactor = th!=1.0;
//遍历vpMapPoints
for(size_t iMP=0; iMP<vpMapPoints.size(); iMP++)
{
MapPoint* pMP = vpMapPoints[iMP];
if(!pMP->mbTrackInView)
continue;
if(pMP->isBad())
continue;
const int &nPredictedLevel = pMP->mnTrackScaleLevel;
// The size of the window will depend on the viewing direction
//根据观测角的cos值确定搜索区域的半径
float r = RadiusByViewingCos(pMP->mTrackViewCos);
if(bFactor)
r*=th;
//pMP通过区域搜索得到F中特征点集合
const vector<size_t> vIndices =
F.GetFeaturesInArea(pMP->mTrackProjX,pMP->mTrackProjY,r*F.mvScaleFactors[nPredictedLevel],nPredictedLevel-1,nPredictedLevel);
if(vIndices.empty())
continue;
const cv::Mat MPdescriptor = pMP->GetDescriptor();
int bestDist=256;
int bestLevel= -1;
int bestDist2=256;
int bestLevel2 = -1;
int bestIdx =-1 ;
// Get best and second matches with near keypoints
//遍历通过区域搜索得到的特征点集合
for(vector<size_t>::const_iterator vit=vIndices.begin(), vend=vIndices.end(); vit!=vend; vit++)
{
const size_t idx = *vit;
//如果这个特征点有对应的mappoint,且其对应的mappoint有被keyframe看到
if(F.mvpMapPoints[idx])
if(F.mvpMapPoints[idx]->Observations()>0)
continue;
if(F.mvuRight[idx]>0)
{
const float er = fabs(pMP->mTrackProjXR-F.mvuRight[idx]);
if(er>r*F.mvScaleFactors[nPredictedLevel])
continue;
}
const cv::Mat &d = F.mDescriptors.row(idx);
const int dist = DescriptorDistance(MPdescriptor,d);
if(dist<bestDist)
{
bestDist2=bestDist;
bestDist=dist;
bestLevel2 = bestLevel;
bestLevel = F.mvKeysUn[idx].octave;
bestIdx=idx;
}
else if(dist<bestDist2)
{
bestLevel2 = F.mvKeysUn[idx].octave;
bestDist2=dist;
}
}
// Apply ratio to second match (only if best and second are in the same scale level)
if(bestDist<=TH_HIGH)
{
if(bestLevel==bestLevel2 && bestDist>mfNNratio*bestDist2)
continue;
F.mvpMapPoints[bestIdx]=pMP;
nmatches++;
}
}
return nmatches;
}
3.2.SearchByProjection()
/**根据上一帧LastFrame的特征点以及所对应的mappoint信息,寻找当前帧的哪些特征点与哪些mappoint的匹配联系
* 根据上一帧特征点对应的3D点投影的位置缩小特征点匹配范围
* @param CurrentFrame 当前帧
* @param LastFrame 上一帧
* @param th 控制特征搜索框的大小阈值
* @param bMono 是否为单目
* @return 成功匹配的数量
*/
int ORBmatcher::SearchByProjection(Frame &CurrentFrame, const Frame &LastFrame, const float th, const bool bMono)
{