Abstract

Many Computer Vision problems arise from information processing of data sources with nuisance variances like scale, orientation, contrast, perspective foreshortening or - in medical imaging - staining and local warping. In most cases these variances can be stated a priori and can be used to improve the generalization of recognition algorithms. We propose a novel supervised feature learning approach, which efficiently extracts information from these constraints to produce interpretable, transformation-invariant features. The proposed method can incorporate a large class of transformations, e.g., shifts, rotations, change of scale, morphological operations, non-linear distortions, photometric transformations, etc. These features boost the discrimination power of a novel image classification and segmentation method, which we call Transformation-Invariant Convolutional Jungles (TICJ). We test the algorithm on two benchmarks in face recognition and medical imaging, where it achieves state of the art results, while being computationally significantly more efficient than Deep Neural Networks.

Keywords

Artificial intelligenceComputer scienceConvolutional neural networkImage warpingPattern recognition (psychology)Invariant (physics)Transformation (genetics)Computer visionFeature extractionSegmentationA priori and a posterioriDeep learningGeneralizationMathematics

Related Publications

Publication Info

Year
2015
Type
article
Pages
3043-3051
Citations
26
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

26
OpenAlex

Cite This

Dmitry Laptev, Joachim M. Buhmann (2015). Transformation-Invariant Convolutional Jungles. , 3043-3051. https://doi.org/10.1109/cvpr.2015.7298923

Identifiers

DOI
10.1109/cvpr.2015.7298923