Abstract

The most successful 2D object detection methods require a large number of images annotated with object bounding boxes to be collected for training. We present an alternative approach that trains on virtual data rendered from 3D models, avoiding the need for manual labeling. Growing demand for virtual reality applications is quickly bringing about an abundance of available 3D models for a large variety of object categories. While mainstream use of 3D models in vision has focused on predicting the 3D pose of objects, we investigate the use of such freely available 3D models for multicategory 2D object detection. To address the issue of dataset bias that arises from training on virtual data and testing on real images, we propose a simple and fast adaptation approach based on decorrelated features. We also compare two kinds of virtual data, one rendered with real-image textures and one without. Evaluation on a benchmark domain adaptation dataset demonstrates that our method performs comparably to existing methods trained on large-scale real image domains.

Keywords

Computer scienceObject detectionBenchmark (surveying)Virtual realityArtificial intelligenceAdaptation (eye)Object (grammar)Computer visionMinimum bounding boxDomain (mathematical analysis)Bounding overwatchVirtual imageMachine learningImage (mathematics)Pattern recognition (psychology)

Affiliated Institutions

Related Publications

Publication Info

Year
2014
Type
article
Pages
82.1-82.12
Citations
167
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

167
OpenAlex

Cite This

Baochen Sun, Kate Saenko (2014). From Virtual to Reality: Fast Adaptation of Virtual Object Detectors to Real Domains. , 82.1-82.12. https://doi.org/10.5244/c.28.82

Identifiers

DOI
10.5244/c.28.82