首页 > 技术文章 > BoW算法及DBoW2库简介(二)

mafuqiang 原文

  一、BoW算法

  用OpenCV实现了最简单的BoW算法进行了一次小规模的图像检索任务,使用UKbench数据库,算法原理和网上的描述差不多,使用K-means算法进行聚类,这里使用KDTree算法进行特征量化,按照自己的理解计算了TF-IDF权重,使用余弦距离计算图像之间的相似性。下面给出关键函数依赖于OpenCV的实现:

如TF-IDF权重的计算,这里只是按照自己的理解实现了算法,有的地方传参不是很合适,不过不影响效果:

std::vector<double> compute_TF(cv::Mat& descriptors, cv::Mat& labels)
{
    std::vector<double> tf(Num_clu, 0.0);
    for (int i = 0; i < descriptors.rows; i++)
    {
        tf[labels.at<int>(i)] ++;
    }

    for (unsigned int i = 0; i < tf.size(); i++)
    {
        tf[i] /= (float)descriptors.rows;
    }
    return tf;
}

std::vector<double> comput_IDF(std::vector<cv::Mat>& descriptors, std::vector<cv::Mat> &labels)
{
    std::vector<double> idf(Num_clu, 1.0);

    for (unsigned int i = 0; i < descriptors.size(); i++)
    {
        std::vector<int> idf_tmp(Num_clu, 0);
        for (int j = 0; j < descriptors[i].rows; j++)
        {
            idf_tmp[labels[i].at<int>(j)] ++;
        }
        for (unsigned int j = 0; j < idf_tmp.size(); j++)
        {
            if (idf_tmp[j] != 0) idf[j] ++;
        }
    }

    for (unsigned int i = 0; i < idf.size(); i++)
    {
        idf[i] = log(Num_img / idf[i]);
    }

    return idf;
}

  有一点需要注意,这里的IDF应该是只计算一次,而TF则是对每一幅图像计算一次。

  有了TF-IDF函数的实现就可以计算BoW向量了,首先是计算训练图像的BoW向量:

cv::Mat TrainingBowVector(cv::Mat & centers, std::vector<double>& IDF)
{
    cv::SurfFeatureDetector detector;
    cv::SurfDescriptorExtractor extractor;

    char image_name[50];
    std::vector<cv::Mat> descriptor_all;
    descriptor_all.reserve(Num_img);

    //Find the keypoints and compute the descriptors;
    for (int i = 1; i <= Num_img; i++)
    {
        std::cout << "I:" << i << std::endl;
        sprintf_s(image_name, "D:\DataBase\UKbench\TestImage\%d.jpg", i);
        cv::Mat image = cv::imread(image_name, 0);
        std::vector<cv::KeyPoint> keypoints;
        cv::Mat descriptors;
        detector.detect(image, keypoints);
        std::cout << "Keypoints:" << keypoints.size() << std::endl;
        extractor.compute(image, keypoints, descriptors);
        descriptor_all.push_back(descriptors);
    }

    //Get the training descriptors;
    std::cout << "Get the training descriptors." << std::endl;
    cv::Mat descriptor_train;
    for (int j = 0; j < Num_tra; j++)
        descriptor_train.push_back(descriptor_all[j]);

    cv::Mat labels_k;
    cv::kmeans(descriptor_train, Num_clu, labels_k, cv::TermCriteria(CV_TERMCRIT_EPS + CV_TERMCRIT_ITER, 100, 0.01)
        , 3, cv::KMEANS_PP_CENTERS, centers);

    const int tk = 1, Emax = INT_MAX;
    cv::KDTree T(centers, false);
    std::vector<cv::Mat> labels(Num_img);
    for (int i = 0; i < Num_img; i++)
    {
        cv::Mat descriptor_img = descriptor_all[i];
        for (int j = 0; j < descriptor_img.rows; j++)
        {
            std::vector<float> desc_vec(descriptor_img.row(j));
            std::vector<int> idx_tmp(tk);
            T.findNearest(desc_vec, tk, Emax, idx_tmp, cv::noArray(), cv::noArray());
            labels[i].push_back(idx_tmp[0]);
        }
    }

    std::cout << "Compute the TF-IDF." << std::endl;
    cv::Mat BowVec;
    //Compute the TF-IDF for each image;
    IDF = comput_IDF(descriptor_all, labels);
    for (int i = 0; i < Num_img; i++)
    {
        std::vector<double> TF = compute_TF(descriptor_all[i], labels[i]);
        cv::Mat BowVec_tmp;
        for (unsigned int j = 0; j < IDF.size(); j++)
        {
            BowVec_tmp.push_back(TF[j] * IDF[j]);
            //BowVec_tmp.push_back(TF[j]);
        }
        BowVec_tmp = BowVec_tmp.t();
        cv::normalize(BowVec_tmp, BowVec_tmp);
        BowVec.push_back(BowVec_tmp);
    }
    return BowVec;
}

  计算测试图片的BoW向量和上面类似。有了训练图像和测试图像的BoW向量就可以根据余弦距离计算相似度了,最后使用堆排序获得最相似的图像ID。

  而Vocabuliary Tree算法的代码实现和上面的不同点在于码书的训练方式。

二、DBoW2库的使用

  使用DBoW2库训练码书,并根据bow打分完成图像检索,根据正向索引完成特征匹配,在ORB里面没注意到倒排索引加速图像检索的部分。

  首先是码书的训练(“盗用”代码:http://www.cnblogs.com/jian-li/p/5666556.html):

#include <iostream>
#include <vector>
#include "Thirdparty/DBoW2/DBoW2/FORB.h"
#include "Thirdparty/DBoW2/DBoW2/TemplatedVocabulary.h"

// OpenCV
#include <opencv2/opencv.hpp>
#include "opencv2/core/core.hpp"
#include <opencv/cv.h>
#include <opencv/highgui.h>
#include <opencv2/nonfree/features2d.hpp>

// ROS
#include <rosbag/bag.h>
#include <rosbag/view.h>
#include <ros/ros.h>
#include <sensor_msgs/Image.h>
#include <boost/foreach.hpp>
#include <cv_bridge/cv_bridge.h>
#include "ORBextractor.h"
#include <dirent.h>
#include <string.h>

using namespace DBoW2;
using namespace DUtils;
using namespace std;
using namespace ORB_SLAM;
// - - - - - --- - - - -- - - - - -

/// ORB Vocabulary
typedef DBoW2::TemplatedVocabulary<DBoW2::FORB::TDescriptor, DBoW2::FORB>
ORBVocabulary;
// - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

void extractORBFeatures(cv::Mat &image, vector<vector<cv::Mat> > &features, ORBextractor* extractor);
void changeStructureORB( const cv::Mat &descriptor,vector<bool> &mask, vector<cv::Mat> &out);
void isInImage(vector<cv::KeyPoint> &keys, float &cx, float &cy, float &rMin, float &rMax, vector<bool> &mask);
void createVocabularyFile(ORBVocabulary &voc, std::string &fileName, const vector<vector<cv::Mat> > &features);
// ----------------------------------------------------------------------------

int main()
{
    //Extracting ORB features from image folder
    vector<std::string> filenames;
    std::string folder = "/home/saodiseng/FRONTAL/";
    cv::glob(folder, filenames);

    // initialze     ORBextractor
    int nLevels = 5;//6;
    ORBextractor* extractor = new ORBextractor(1000,1.2,nLevels,1,20);
    int nImages = filenames.size();

    vector<vector<cv::Mat > > features;
    features.clear();
    features.reserve(nImages);

    cv::Mat image;
    cout << "> Extracting Features from " << nImages << " images..." << endl;
    for(int i = 0; i < nImages; ++i)
    {
        std::cout << "Processing the " << i <<" image " << std::endl;
        cv::Mat src = cv::imread(filenames[i]);
        imshow("View", src);
        cv::waitKey(1);
        if (!src.empty())
        {
            cv::cvtColor(src, image, CV_RGB2GRAY);
            extractORBFeatures(image, features, extractor);
        }
    }
    cout << "... Extraction done!" << endl;

    // Creating the Vocabulary
    // define vocabulary
    const int k = 10; // branching factor
    const WeightingType weight = TF_IDF;
    const ScoringType score = L1_NORM;
    ORBVocabulary voc(k, nLevels, weight, score);
    std::string vociName = "vociOmni.txt";
    createVocabularyFile(voc, vociName, features);
    cout << "--- THE END ---" << endl;
    
    return 0;
}
// ----------------------------------------------------------------------------

void extractORBFeatures(cv::Mat &image, vector<vector<cv::Mat> > &features, ORBextractor* extractor) {
    vector<cv::KeyPoint> keypoints;
    cv::Mat descriptorORB;
    (*extractor)(image, cv::Mat(), keypoints, descriptorORB);

    // reject features outside region of interest
    vector<bool> mask;
    float cx = 0; float cy = 0;
    float rMin = 0; float rMax = 0;
    isInImage(keypoints, cx, cy, rMin, rMax, mask);
    // create descriptor vector for the vocabulary
    features.push_back(vector<cv::Mat>());
    changeStructureORB(descriptorORB, mask, features.back());
    imshow("ORBFeature", features.back().back());
}

// ----------------------------------------------------------------------------

void changeStructureORB( const cv::Mat &descriptor,vector<bool> &mask, vector<cv::Mat> &out) {
    for (int i = 0; i < descriptor.rows; i++) {
        if(mask[i]) {
            out.push_back(descriptor.row(i));
        }
    }
}

// ----------------------------------------------------------------------------

void isInImage(vector<cv::KeyPoint> &keys, float &cx, float &cy, float &rMin, float &rMax, vector<bool> &mask) {
    int N = keys.size();
    mask = vector<bool>(N, false);
    int num = 0;
    for(int i=0; i<N; i++) {
        cv::KeyPoint kp = keys[i];
        float u = kp.pt.x;
        float v = kp.pt.y;
        if(u>20 && u<320-20 && v>20 && v<240-20)
        {
            mask[i] = true;
            num ++;
        }
    }
    std::cout << "In image number " << num << std::endl;
}

// ----------------------------------------------------------------------------

void createVocabularyFile(ORBVocabulary &voc, std::string &fileName, const vector<vector<cv::Mat> > &features)
{
    cout << "> Creating vocabulary. May take some time ..." << endl;
    voc.create(features);
    cout << "... done!" << endl;
    cout << "> Vocabulary information: " << endl
        << voc << endl << endl;
    // save the vocabulary to disk
    cout << endl << "> Saving vocabulary..." << endl;
    voc.saveToTextFile(fileName);
    cout << "... saved to file: " << fileName << endl;
}

   也可以直接使用ORB给定的码书。

  再下面就是训练BoW向量并计算打分:

void FrameRecog::ComputeBoW()
{
    //数据类型转换;
    vector<cv::Mat>vFrDesc = Converter::toDescriptorVector(Descriptors);
    //BowVec为BoW特征向量,FeatVec为正向索引;
    pORBVocabulary->transform(vFrDesc, BowVec, FeatVec, 4);
}



float score = pORBVocabulary->score(BowVec, vBowVec[i]);

ComputeBoW()函数计算了当前帧的BowVec向量,以及它的第4层正向索引值FeatVec。下面一句即计算了两个BoW向量的相似性打分。当打分满足某个阈值之后,还需要通过正向索引值进行特征匹配:

int FrameRecog::FeatMatchByBoW( const int idx )
{
    int nmatches = 0;
    const int TH_LOW = 50;
    const int HISTO_LENGTH = 30;
    const int factor = 1.0f/HISTO_LENGTH;

    const DBoW2::FeatureVector &vFeatVecTD = vFeatVec[idx];
    const DBoW2::FeatureVector &vFeatVecCD = FeatVec; 

    DBoW2::FeatureVector::const_iterator TDit = vFeatVecTD.begin();
    DBoW2::FeatureVector::const_iterator CDit = vFeatVecCD.begin();
    DBoW2::FeatureVector::const_iterator TDend= vFeatVecTD.end();
    DBoW2::FeatureVector::const_iterator CDend= vFeatVecCD.end();

    while( TDit != TDend && CDit != CDend )
    {
               //first为单词的索引,second则对应为该单词索引下的ORB特征集合;
        if( TDit->first == CDit->first)
        {
        //second是要循环的对象
            const vector<unsigned int> vIndicesTD = TDit->second;
            const vector<unsigned int> vIndicesCD = CDit->second;
              
        //循环关键帧和当前帧对应单词下的特征集合,计算相似性;
            for ( size_t iTD = 0; iTD < vIndicesTD.size(); iTD ++ )
            {
                const unsigned int realIdxTD = vIndicesTD[iTD];
                const cv::Mat &dTD = vDescriptors[idx].row(realIdxTD);

                int bestDist1 = 256;
                int bestIdxF  = -1;
                int bestDist2 = 256;

                for ( size_t iCD = 0; iCD < vIndicesCD.size(); iCD ++ )
                {
                    const unsigned int realIdxCD = vIndicesCD[iCD];
                    const cv::Mat &dCD = Descriptors.row(realIdxCD);
                    const int dist = DescriptorDistance(dTD, dCD);
             //这里注意是双阈值;
                    if( dist < bestDist1 )
                    {
                        bestDist2 = bestDist1;
                        bestDist1 = dist;
                        bestIdxF  = realIdxCD;
                    }
                    else if( dist < bestDist2 )
                    {
                        bestDist2 = dist;
                    }
                }

           //这里有两个输入参数,一个是TH_LOW,是指两个特征的最小距离阈值;
           //第二个是0.95,它是指相似特征的最小距离小于第二小距离的百分之九十五;
           //第二个参数的含义是,当该参数越接近于1时,该式越接近于成立,而越小时说明要求越高,
           //即最小距离远大于第二小距离,所以两特征是相似特征的概率非常大
                if(bestDist1 <= TH_LOW)
                {
                    if( static_cast<float>(bestDist1)<0.95 * static_cast<float>(bestDist2))
                        nmatches ++;
                }
            }
            TDit ++;
            CDit ++;
        }
        else if( TDit->first < CDit->first )
        {
            TDit = vFeatVecTD.lower_bound(CDit->first);
        }
        else
        {
            CDit = vFeatVecCD.lower_bound(TDit->first);
        }
    }

   //原函数中还有特征对应的3D地图点的输出,以及根据ORB特征的主方向进一步判断特征是否相似的代码,这里略去;
    return nmatches;
}
int FrameRecog::DescriptorDistance(const cv::Mat &a, const cv::Mat &b)
{
    const int *pa = a.ptr<int32_t>();
    const int *pb = b.ptr<int32_t>();

    int dist = 0;
    for ( int i = 0; i < 8; i ++, pa ++, pb ++ )
    {
        unsigned int v = *pa ^ *pb;
        v = v - ((v>>1) & 0x55555555);
        v = (v & 0x33333333) + ((v >> 2) & 0x33333333);
        dist += (((v + (v >> 4)) & 0xF0F0F0F) * 0x1010101) >> 24;
    }
    return dist;
}

 

上面的源文件在ORBmatches.cc中的

int ORBmatcher::SearchByBoW(KeyFrame* pKF,Frame &F, vector<MapPoint*> &vpMapPointMatches) 函数中。即通过正向索引给出特征匹配数或匹配的特征以及对应的3D点。

推荐阅读