Abstract
Three-dimensional shapes contain different kinds of information that jointly characterize the shape. Traditional methods, however, perform recognition or retrieval using only one type. This article presents a 3D feature learning framework that combines different modality data effectively to promote the discriminability of unimodal features. Two independent deep belief networks (DBNs) are employed to learn high-level features from low-level features, and a restricted Boltzmann machine (RBM) is trained for mining the deep correlations between the different modalities. Experiments demonstrate that the proposed method can achieve better performance.
Original language | English |
---|---|
Article number | 52 |
Pages (from-to) | 38-46 |
Number of pages | 9 |
Journal | IEEE Multimedia |
Volume | 21 |
Issue number | 4 |
DOIs | |
State | Published - 1 Oct 2014 |
Keywords
- Accuracy
- Deep learning
- Feature extraction
- Fusion
- Learning systems
- Multimedia
- Multimodal feature fusion
- Research and development
- Shape analysis
- Shape recognition
- Shape retrieval
- Solid modeling
- Three-dimensional displays