|
|
|
|
LEADER |
05690nma a2201405 u 4500 |
001 |
EB002143049 |
003 |
EBX01000000000000001281175 |
005 |
00000000000000.0 |
007 |
cr||||||||||||||||||||| |
008 |
230202 ||| eng |
020 |
|
|
|a books978-3-0365-6062-5
|
020 |
|
|
|a 9783036560625
|
020 |
|
|
|a 9783036560618
|
100 |
1 |
|
|a Lv, Zhihan
|
245 |
0 |
0 |
|a Augmented Reality, Virtual Reality & Semantic 3D Reconstruction
|h Elektronische Ressource
|
260 |
|
|
|a Basel
|b MDPI - Multidisciplinary Digital Publishing Institute
|c 2022
|
300 |
|
|
|a 1 electronic resource (304 p.)
|
653 |
|
|
|a continuous performance test
|
653 |
|
|
|a one-shot learning
|
653 |
|
|
|a assessment
|
653 |
|
|
|a algorithm
|
653 |
|
|
|a skeleton-based
|
653 |
|
|
|a deep learning
|
653 |
|
|
|a super-resolution
|
653 |
|
|
|a 360° video
|
653 |
|
|
|a positioning
|
653 |
|
|
|a History of engineering & technology / bicssc
|
653 |
|
|
|a empowerment
|
653 |
|
|
|a multi-view stereo
|
653 |
|
|
|a local feature
|
653 |
|
|
|a Technology: general issues / bicssc
|
653 |
|
|
|a EDAH
|
653 |
|
|
|a bibliometric analysis
|
653 |
|
|
|a texture loss
|
653 |
|
|
|a audio classification
|
653 |
|
|
|a cooperative games
|
653 |
|
|
|a virtual reality (VR)
|
653 |
|
|
|a scientific production
|
653 |
|
|
|a pre-visualization
|
653 |
|
|
|a olfactory display
|
653 |
|
|
|a dense convolutional networks
|
653 |
|
|
|a game engine
|
653 |
|
|
|a WGAN-GP
|
653 |
|
|
|a projection mapping
|
653 |
|
|
|a 3D composition
|
653 |
|
|
|a GREN
|
653 |
|
|
|a computer vision
|
653 |
|
|
|a anatomization
|
653 |
|
|
|a problem-solving
|
653 |
|
|
|a motor planning
|
653 |
|
|
|a 3P model
|
653 |
|
|
|a 3D face model
|
653 |
|
|
|a three-dimensional reconstruction
|
653 |
|
|
|a semantic 3D reconstruction
|
653 |
|
|
|a ADHD
|
653 |
|
|
|a augmented reality
|
653 |
|
|
|a web of science
|
653 |
|
|
|a image matching
|
653 |
|
|
|a 3D reconstruction
|
653 |
|
|
|a fully convolutional network (FCN)
|
653 |
|
|
|a mobile lip reading system
|
653 |
|
|
|a graph-based refinement
|
653 |
|
|
|a robotic manipulator
|
653 |
|
|
|a panoramic photography
|
653 |
|
|
|a stereo vision
|
653 |
|
|
|a higher education
|
653 |
|
|
|a spatial information
|
653 |
|
|
|a n/a
|
653 |
|
|
|a transformation
|
653 |
|
|
|a interactive learning environments
|
653 |
|
|
|a area of interest
|
653 |
|
|
|a orientation
|
653 |
|
|
|a generative adversarial networks
|
653 |
|
|
|a super-resolution reconstruction
|
653 |
|
|
|a photoreality
|
653 |
|
|
|a superpixel
|
653 |
|
|
|a Photometric Stereo (PS)
|
653 |
|
|
|a children
|
653 |
|
|
|a virtual reality
|
653 |
|
|
|a educational technology
|
653 |
|
|
|a viewpoint
|
653 |
|
|
|a transfer learning
|
653 |
|
|
|a Dynamic Time Warping
|
653 |
|
|
|a perception
|
653 |
|
|
|a applications in subject areas
|
653 |
|
|
|a lightweight neural network
|
653 |
|
|
|a scientific mapping
|
653 |
|
|
|a laser scanning
|
653 |
|
|
|a 3D modelling
|
653 |
|
|
|a primary education
|
653 |
|
|
|a construction hazard
|
653 |
|
|
|a feature tracking
|
653 |
|
|
|a face correction
|
653 |
|
|
|a gesture recognition
|
653 |
|
|
|a radial curve
|
653 |
|
|
|a probabilistic fusion
|
653 |
|
|
|a eye-in-hand vision system
|
653 |
|
|
|a wayfinding
|
653 |
|
|
|a safety education
|
653 |
|
|
|a 3D representation
|
653 |
|
|
|a inception model
|
653 |
|
|
|a semi-immersive virtual reality
|
653 |
|
|
|a structure from motion
|
700 |
1 |
|
|a Wang, Jing-Yan
|
700 |
1 |
|
|a Kumar, Neeraj
|
700 |
1 |
|
|a Lloret, Jaime
|
041 |
0 |
7 |
|a eng
|2 ISO 639-2
|
989 |
|
|
|b DOAB
|a Directory of Open Access Books
|
500 |
|
|
|a Creative Commons (cc), https://creativecommons.org/licenses/by/4.0/
|
028 |
5 |
0 |
|a 10.3390/books978-3-0365-6062-5
|
856 |
4 |
0 |
|u https://www.mdpi.com/books/pdfview/book/6481
|7 0
|x Verlag
|3 Volltext
|
856 |
4 |
2 |
|u https://directory.doabooks.org/handle/20.500.12854/95825
|z DOAB: description of the publication
|
082 |
0 |
|
|a 363
|
082 |
0 |
|
|a 000
|
082 |
0 |
|
|a 370
|
082 |
0 |
|
|a 900
|
082 |
0 |
|
|a 770
|
082 |
0 |
|
|a 600
|
082 |
0 |
|
|a 620
|
520 |
|
|
|a Augmented reality is a key technology that will facilitate a major paradigm shift in the way users interact with data and has only just recently been recognized as a viable solution for solving many critical needs. In practical terms, this innovation can be used to visualize data from hundreds of sensors simultaneously, overlaying relevant and actionable information over your environment through a headset. Semantic 3D reconstruction unlocks the promise of AR technology, possessing a far greater availability of semantic information. Although, there are several methods currently available as post-processing approaches to extract semantic information from the reconstructed 3D models, the results obtained results have been uncertain and evenly incorrect. Thus, it is necessary to explore or develop a novel 3D reconstruction approach to automatically recover 3D geometry model and obtained semantic information simultaneously. The rapid advent of deep learning brought new opportunities to the field of semantic 3D reconstruction from photo collections. Deep learning-based methods are not only able to extract semantic information but can also enhance fundamental techniques in semantic 3D reconstruction, techniques which include feature matching or tracking, stereo matching, camera pose estimation, and use of multi-view stereo methods. Moreover, deep learning techniques can be used to extract priors from photo collections, and this obtained information can in turn improve the quality of 3D reconstruction.
|