Rethinking Image Cropping: Exploring Diverse Compositions From Global Views | |
Jia, Gengyun1,2; Huang, Huaibo1,2; Fu, Chaoyou1,2; He, Ran1,2 | |
2022 | |
会议名称 | IEEE Conference on Computer Vision and Pattern Recognition |
会议日期 | 2022.6.19 |
会议地点 | 路易斯安那新奥尔良 |
摘要 | Existing image cropping works mainly use anchor evaluation methods or coordinate regression methods. However, it is difficult for pre-defined anchors to cover good crops globally, and the regression methods ignore the cropping diversity. In this paper, we regard image cropping as a set prediction problem. A set of crops regressed from multiple learnable anchors is matched with the labeled good crops, and a classifier is trained using the matching results to select a valid subset from all the predictions. This new perspective equips our model with globality and diversity, mitigating the shortcomings but inherit the strengthens of previous methods. Despite the advantages, the set prediction method causes inconsistency between the validity labels and the crops. To deal with this problem, we propose to smooth the validity labels with two different methods. The first method that uses crop qualities as direct guidance is designed for the datasets with nearly dense quality labels. The second method based on the self distillation can be used in sparsely labeled datasets. Experimental results on the public datasets show the merits of our approach over state- of-the-art counterparts. |
文献类型 | 会议论文 |
条目标识符 | http://ir.ia.ac.cn/handle/173211/48683 |
专题 | 模式识别实验室 |
通讯作者 | He, Ran |
作者单位 | 1.School of Artificial Intelligence, University of Chinese Academy of Sciences 2.NLPR & CRIPAC, Institute of Automation, Chinese Academy of Sciences |
第一作者单位 | 模式识别国家重点实验室 |
通讯作者单位 | 模式识别国家重点实验室 |
推荐引用方式 GB/T 7714 | Jia, Gengyun,Huang, Huaibo,Fu, Chaoyou,et al. Rethinking Image Cropping: Exploring Diverse Compositions From Global Views[C],2022. |
条目包含的文件 | 下载所有文件 | |||||
文件名称/大小 | 文献类型 | 版本类型 | 开放类型 | 使用许可 | ||
CVPR_JIA.pdf(5469KB) | 会议论文 | 开放获取 | CC BY-NC-SA | 浏览 下载 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论