CASIA OpenIR  > 毕业生  > 博士学位论文
Thesis Advisor吴福朝
Degree Grantor中国科学院大学
Place of Conferral北京
Keyword机器学习 度量学习 非欧几何度量 凯莱-克莱因度量
    1. 根据样本数据的统计特征(均值、协方差矩阵)和凯莱-克莱因度量的定义,构造了一种广义马氏度量,并指出了广义马氏度量与马氏度量之间的关系。在机器学习数据库中的性能测试表明:广义马氏度量比欧氏度量、传统的马氏度量有更好的分类性能。
    2. 提出一种基于MMC(Mahalanobis Metric Learning for Clustering,聚类马氏度量学习)准则的凯莱-克莱因度量学习算法(CK-MMC)。应用MMC的学习准则得到一个凯莱-克莱因度量,在最小化相似样本之间的凯莱-克莱因度量距离的同时最大化不相似样本之间的凯莱-克莱因度量距离。在CK-MMC问题中,使用梯度上升算法进行求解,实验结果表明CK-MMC有效地提升了传统MMC度量学习算法的分类性能。
    3. 提出一种基于LMNN(Large Margin Nearest Neighbors,最大间隔近邻)准则的凯莱-克莱因度量学习算法(CK-LMNN)。CK-LMNN使用相对距离约束学习得到一个凯莱-克莱因度量,使得每个样本与它的k-近邻同类,同时与不同类的样本点保持较大的距离。CK-LMNN问题使用梯度下降算法求解,实验结果表明CK-LMNN有效地提升了传统LMNN的分类性能。
    4. 提出一种多凯莱-克莱因度量学习算法(MCKML)。由同类样本集构造一个局部凯莱-克莱因度量,再将所有局部度量进行线性组合得到样本空间上的度量,称为多凯莱-克莱因度量。由于凯莱-克莱因度量内在的非线性,多凯莱-克莱因度量可以更好地模拟数据空间的非线性结构。在MCKML中,使用局部度量和组合系数交替优化的方式求解,实验结果表明使用多凯莱-克莱因度量有效提升了单个凯莱-克莱因度量学习算法的分类性能。
    5. 提出一种自适应约束凯莱-克莱因度量学习算法(CKseML)。在原始凯莱-克莱因度量学习算法的目标函数中,使用固定边界生成约束条件,对空间分布较为复杂的样本适应性不好。CKseML引入自适应收缩扩张约束,根据当前样本空间的分布状态,适时地收缩/扩张成对约束,以松弛边界约束条件,从而能适应分布更为复杂的样本空间。CKseML使用交替迭代方法优化凯莱-克莱因度量矩阵,实验结果表明CKseML比固定边界约束的凯莱-克莱因度量学习算法有更好的分类性能。
Other Abstract
    Distance metric learning plays an important role in many computer vision and pattern recognition tasks, such as image recognition, retrieval and clustering. Existing research in metric learning includes linear and non-linear approaches. Linear metric learning methods aim to learn a proper Mahalanobis metric to minimize the given objective function. Non-linear metric learning methods obtain a more generalized metric by means of kernel trick, manifold learning and deep learning. In this paper, we present a non-linear metric learning framework based on Cayley-Klein metric, aiming to improve the performance of the traditional Mahalanobis metric learning method when dealing with complex data space. Our main contributions include:
    1. Based on the statistical properties of data space (mean and covariance matrix) and the definition of Cayley-Klein metric, we propose a special form of Cayley-Klein metric called generalized Mahalanobis metric and discuss its relationship with Mahalanobis metric. Experimental results show that the generalized Mahalanobis metric outperforms both Euclidean and traditional Mahalanobis metrics in classification task.
    2. We propose a Cayley-Klein metric learning method (CK-MMC) based on MMC (Mahalanobis Metric Learning for Clustering). CK-MMC aims to minimize distances between similar pairs and simultaneously maximize distances between dissimilar pairs. The optimization problem can be solved by the gradient ascent algorithm. Experiments have shown the superiority of CK-MMC over the traditional MMC metric learning method.
    3. We propose a Cayley-Klein metric learning method (CK-LMNN) based on LMNN (Large Margin Nearest Neighbors). CK-LMNN utilizes triplet-wise constraints to learn a Cayley-Klein metric, aiming to make the k-nearest neighbors of a data point lie in the same class as the data point, and meanwhile make data points from different classes are separated by a large margin. The optimization problem can be solved by the gradient descent algorithm. Experiments have shown the superiority of CK-LMNN over the traditional LMNN metric learning method.
    4. We propose a multiple Cayley-Klein metric learning method (MCKML). Multiple Cayley-Klein metric is defined as a linear combination of several local Cayley-Klein metrics, which are learned by data points from the same class. Since Cayley-Klein metric is a kind of non-linear metric, its combination could model the data space better, thus lead to an improved performance. MCKML is learned by iterative optimization over single Cayley-Klein metrics and their combination coefficients. Experiments on several benchmarks show that MCKML outperforms the single Cayley-Klein metric learning method.
    5. We propose an adaptive constraints based Cayley-Klein metric learning method (CKseML). The original Cayley-Klein metric learning methods use conventional pairwise and triplet-wise constraints, which are fixed bound based constraints, may not perform well when the data distribution become complex. CKseML adaptively shrinks/expands the pairwise constraints considering the current data distribution, leading to better adaptability for complex data distribution. Experimental results demonstrate that CKseML achieves better performance than the fixed constraints based Cayley-Klein metric learning methods.
    This paper first introduces the Cayley-Klein metric in the community of metric learning. As a special kind of Riemannian metric learning, it is a powerful alternative to the traditional Mahalanobis metric and can be widely used in computer vision and pattern recognition.
Document Type学位论文
First Author AffilicationInstitute of Automation, Chinese Academy of Sciences
Recommended Citation
GB/T 7714
毕琰虹. 凯莱-克莱因度量学习[D]. 北京. 中国科学院大学,2018.
Files in This Item:
File Name/Size DocType Version Access License
Thesis_byh.pdf(4575KB)学位论文 暂不开放CC BY-NC-SAApplication Full Text
Related Services
Recommend this item
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[毕琰虹]'s Articles
Baidu academic
Similar articles in Baidu academic
[毕琰虹]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[毕琰虹]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.