ICode9

精准搜索请尝试: 精确搜索
首页 > 其他分享> 文章详细

『论文笔记』A Benchmark on Tricks for Large-scale Image Retrieval

2021-10-23 16:34:59  阅读:273  来源: 互联网

标签:training scale set Benchmark Tricks GLD virtual classes TR1


2. Pre-processing Tricks

2.1. Dataset Cleaning for Training

TR1

We noticed by visual inspection that the training set of GLD v1 is clean and reliable, so we used the dataset as is for training. To obtain a semi-supervised learning effect [4, 23], we added virtual classes to the training set. These virtual classes are the clusters from the test and index sets of GLD v1. First, we trained a baseline model using GLD v1 and extracted features of the test and index set of GLD v1. Then, DBSCAN was applied to generate clusters, where each cluster was assigned as a new virtual classes. We call the result TR1 for clarity.

TR2

GLD2的index set中包含很多非地标图片,为了训练模型更好的区分地标和非地标,作者设计了TR2。

和TR1类似,训练后进行聚类,然后“picked several distractor clusters as virtual classes”,将这些干扰标签和TR1的虚拟标签合并,作为TR2。

TR3

The training set of GLD v2 has more classes and images than GLD v1 does,同时也有更多噪声。先用二分类模型移除自然景观图片,然后对每个类别内部进行聚类,当一个类内有多个cluster时,保留最大的cluster。

(同时排除了和TR2重复的类别?)

2.2. Small-scale Validation

从训练集合中抽取2%,分别为测试集合和索引集合。然后包含一个噪音类的虚拟簇(We included a virtual class from a noise cluster),这是因为GLD v2包含有噪音,这样分布能够更加的贴合数据集。

2.3. Experimental Results

作者的目的就是剔出下图中展示的地标图的局部照片。

 

原始数据集的类内问题:the raw dataset includes images taken from inside, outside, and even partial viewpoints from within the same landmark.These kinds of datasets with large intra-class variation may interfere with learning proper representations in the model, especially when a pair-wise ranking loss is used

原始数据集的类间问题:images of nature scenes also make the training process hard as they have a little iter-class variation.

 

 

GLD v2噪声太大,没法直接用于训练。Valid是自建验证集,后面两个指标应该是数据集网站在作者提交模型后给的评分。

Training with TR1, which contains virtual classes from the test and index sets, improves the model performance by using unlabeled data when the original training set is not helpful anymore.

The model trained with TR2 gives performance similar to the model trained with TR1 because the number of data and classes is not noticeably different

Using TR3, which includes a cleaned training set from GLD v2, further improved

 3. Learning Representations

 

 

3.1. Pooling

3.2. Objectives

Xent + Triplet

Triplet loss + classification loss such as cross-entropy (Xent) loss

N-pair + Angular

参见:https://blog.csdn.net/update7/article/details/112391276

3.3. Training a Single Model

标签:training,scale,set,Benchmark,Tricks,GLD,virtual,classes,TR1
来源: https://www.cnblogs.com/hellcat/p/15448789.html

本站声明: 1. iCode9 技术分享网(下文简称本站)提供的所有内容,仅供技术学习、探讨和分享;
2. 关于本站的所有留言、评论、转载及引用,纯属内容发起人的个人观点,与本站观点和立场无关;
3. 关于本站的所有言论和文字,纯属内容发起人的个人观点,与本站观点和立场无关;
4. 本站文章均是网友提供,不完全保证技术分享内容的完整性、准确性、时效性、风险性和版权归属;如您发现该文章侵犯了您的权益,可联系我们第一时间进行删除;
5. 本站为非盈利性的个人网站,所有内容不会用来进行牟利,也不会利用任何形式的广告来间接获益,纯粹是为了广大技术爱好者提供技术内容和技术思想的分享性交流网站。

专注分享技术,共同学习,共同进步。侵权联系[81616952@qq.com]

Copyright (C)ICode9.com, All Rights Reserved.

ICode9版权所有