RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition Presented by Qilong Wang and Peihua Li Dalian University of Technology http://ice.dlut.edu.cn/PeihuaLi/ Kernel SVM Infinite-dimensional Covariance for classification Hand-crafted features Image [23] [20] [17] Feature extraction Feature mapping in RKHS Image model M. T. Harandi, M. Salzmann, and F. M. Porikli. Bregman divergences for infinite dimensional covariance matrices. In CVPR, 2014 Ha Quang Minh, Marco San Biagio and Vittorio Murino. Log-Hilbert-Schmidt metric between positive definite operators on Hilbert spaces. In NIPS, 2014. M. Faraki, M. Harandi, and F. Porikli. Approximate infinitedimensional region covariance descriptors for image classification. In ICASSP, 2015 Comparison of different infinite-dimensional image representation Methods Zhou et al. [PAMI06] Representation Estimator RBF kernel (no explicit mapping) Ledoit-Wolf estimator Covariance Faraki et al. [17] RIAD-G (Ours) Linear SVM? Gaussian Harandi et al. [23] Log-HS [20] Kernels & Mapping Random Fourier transform Nystrom method for RBF kernel Gaussian Hellinger’s kernel 𝒳2 kernel Computationally very expensive and unscalable Feature mapping by RBF kernel & kernel SVM [20,23] ● Inefficient approximate mappings [17] ● Unrobust covariance estimation [17,20,23] No Σ + λI Yes vN-MLE Yes Hard to use very high-dimensional features Proposed RAID-G ̶ Our doings CNN features Hellinger’s kernel Chi-Square kernel Image Feature extraction Analytical, finitedimensional mapping φHel (xk ) = xk Gaussian representation φˆ Chi (xk ) = xk L, 2Lsech(Lπ)cos(Llog(xk )) 2Lsech(Lπ)sin(Llog(xk )) T [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 Proposed RAID-G ~500 features CNN High-dimension & small sample problem I am robust Hellinger’s kernel ~ Chi-Square kernel Image Feature extraction Analytical, finitedimensional mapping φHel (xk ) = xk Robust Gaussian φˆ Chi (xk ) = xk L, 2Lsech(Lπ)cos(Llog(xk )) 2Lsech(Lπ)sin(Llog(xk )) T [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 Proposed RAID-G ̶ robust estimator Classical MLE(not robust in our case) where Robust vN-MLE where is the von Neumann divergence between matrices. [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 Proposed RAID-G ̶ Flowchart Approximate RKHS Mapping . . . ... Robust Estimation ... . . . . . . Mapping Deep CNN features RIAD-G Modeling images with RIAD-G in multiscale manner VGG-VD16 without fine-tuning Gaussian Embedding & vectorization Linear SVM Results on Material Recognition Gaussian representation always performs better than covariance. The vN-MLE estimator significantly improves the classical MLE. Feature mapping by kernel embeddings brings benefits. RAID-G outperforms FV-CNN, achieving state-of-the-art results. [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 Comparison with other robust estimators [LW] O. Ledoit and M. Wolf. A well-conditioned estimator for large-dimensional covariance matrices. J. Multivariate Analysis, 2004. [Stein ] M. J. Daniels and R. E. Kass. Shrinkage estimators for covariance matrices. Biometrics, 2001. [MMSE] Y. Chen, A. Wiesel, et al. Shrinkage algorithms for MMSE covariance estimation. IEEE TSP, 2010. [EL-SP] E. Yang, A. Lozano, et al. Elementary estimators for sparse covariance matrices and other structured moments. In ICML, 2014. [vN-MLE]Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 . Comparison with other explicit kernel mappings [17] M. Faraki, M. Harandi, and F. Porikli. Approximate infinitedimensional region covariance descriptors for image classification. In ICASSP, 2015 [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 Comparison with infinite dimensional covariance on KTH-TIPS 2b [23, 20] are slightly better than RAID-G with hand-crafted features. RAID-G outperforms [23,20] by 7% with CNN features, while [23,20] involves unaffordable cost with CNN features. [23] M. T. Harandi, M. Salzmann, and F. M. Porikli. Bregman divergences for infinite dimensional covariance matrices. In CVPR, 2014 [20] Ha Quang Minh, Marco San Biagio and Vittorio Murino. Log-Hilbert-Schmidt metric between positive definite operators on Hilbert spaces. In NIPS, 2014. [RAID-G] Qilong Wang, Peihua Li, Wangmeng Zuo, and Lei Zhang. RAID-G: Robust Estimation of Approximate Infinite Dimensional Gaussian with Application to Material Recognition, In CVPR, 2016 . Summary RAID-G is a very competitive image representation √ Efficient (analytic feature mapping & covariance computation) √ Scalable (linear SVM) √ Superior performance Robust estimation is critical for higher dimension problem (>512) Analytical, explicit feature mappings improve performance
© Copyright 2026 Paperzz