Ausgabe 2/2024
Inhalt (15 Artikel)
Single Pixel Spectral Color Constancy
Samu Koskinen, Erman Acar, Joni-Kristian Kämäräinen
CDistNet: Perceiving Multi-domain Character Distance for Robust Text Recognition
Tianlun Zheng, Zhineng Chen, Shancheng Fang, Hongtao Xie, Yu-Gang Jiang
Blind Image Deblurring with Unknown Kernel Size and Substantial Noise
Zhong Zhuang, Taihui Li, Hengkang Wang, Ju Sun
A Region-Based Randers Geodesic Approach for Image Segmentation
Da Chen, Jean-Marie Mirebeau, Huazhong Shu, Laurent D. Cohen
Transferring Vision-Language Models for Visual Recognition: A Classifier Perspective
Wenhao Wu, Zhun Sun, Yuxin Song, Jingdong Wang, Wanli Ouyang
Learning Robust Facial Representation From the View of Diversity and Closeness
Chaoyu Zhao, Jianjun Qian, Shumin Zhu, Jin Xie, Jian Yang
Deep Richardson–Lucy Deconvolution for Low-Light Image Deblurring
Liang Chen, Jiawei Zhang, Zhenhua Li, Yunxuan Wei, Faming Fang, Jimmy Ren, Jinshan Pan
Intra- & Extra-Source Exemplar-Based Style Synthesis for Improved Domain Generalization
Yumeng Li, Dan Zhang, Margret Keuper, Anna Khoreva
Sfnet: Faster and Accurate Semantic Segmentation Via Semantic Flow
Xiangtai Li, Jiangning Zhang, Yibo Yang, Guangliang Cheng, Kuiyuan Yang, Yunhai Tong, Dacheng Tao
MineGAN++: Mining Generative Models for Efficient Knowledge Transfer to Limited Data Domains
Yaxing Wang, Abel Gonzalez-Garcia, Chenshen Wu, Luis Herranz, Fahad Shahbaz Khan, Shangling Jui, Jian Yang, Joost van de Weijer
One-Pot Multi-frame Denoising
Lujia Jin, Qing Guo, Shi Zhao, Lei Zhu, Qian Chen, Qiushi Ren, Yanye Lu
The Curious Layperson: Fine-Grained Image Recognition Without Expert Labels
Subhabrata Choudhury, Iro Laina, Christian Rupprecht, Andrea Vedaldi
Building 3D Generative Models from Minimal Data
Skylar Sutherland, Bernhard Egger, Joshua Tenenbaum
CLIP-Adapter: Better Vision-Language Models with Feature Adapters
Peng Gao, Shijie Geng, Renrui Zhang, Teli Ma, Rongyao Fang, Yongfeng Zhang, Hongsheng Li, Yu Qiao
How Does Fine-Tuning Impact Out-of-Distribution Detection for Vision-Language Models?
Yifei Ming, Yixuan Li