[-] Show simple item record

dc.contributor.advisorScott, Grant (Grant J.)eng
dc.contributor.authorYang, Alex (M.S. in computer science)eng
dc.date.issued2017eng
dc.date.submitted2017 Falleng
dc.descriptionField of study: Computer science.eng
dc.descriptionDr. Grant Scott, Thesis Supervisor.eng
dc.description"December 2017."eng
dc.description.abstractDepth estimation from single monocular images is a theoretical challenge in computer vision as well as a computational challenge in practice. This thesis addresses the problem of depth estimation from single monocular images using a deep convolutional neural fields framework; which consists of convolutional feature extraction, superpixel dimensionality reduction, and depth inference. Data were collected using a stereo vision camera, which generated depth maps though triangulation that are paired with visual images. The visual image (input) and computed depth map (desired output) are used to train the model, which has achieved 83 percent test accuracy at the standard 25 percent tolerance. The problem has been formulated as depth regression for superpixels and our technique is superior to existing state-of-the-art approaches based on its demonstrated its generalization ability, high prediction accuracy, and real-time processing capability. We utilize the VGG-16 deep convolutional network as feature extractor and conditional random fields depth inference. We have leveraged a multi-phase training protocol that includes transfer learning and network fine-tuning lead to high performance accuracy. Our framework has a robust modular nature with capability of replacing each component with different implementations for maximum extensibility. Additionally, our GPU-accelerated implementation of superpixel pooling has further facilitated this extensibility by allowing incorporation of feature tensors with exible shapes and has provided both space and time optimization. Based on our novel contributions and high-performance computing methodologies, the model achieves a minimal and optimized design. It is capable of operating at 30 fps; which is a critical step towards empowering real-world applications such as autonomous vehicle with passive relative depth perception using single camera vision-based obstacle avoidance, environment mapping, etc.eng
dc.description.bibrefIncludes bibliographical references (pages 61-65).eng
dc.format.extent1 online resource (viii, 65 pages) : illustrations (chiefly color)eng
dc.identifier.merlinb129592006eng
dc.identifier.oclc1101434765eng
dc.identifier.urihttps://hdl.handle.net/10355/63579
dc.identifier.urihttps://doi.org/10.32469/10355/63579eng
dc.languageEnglisheng
dc.publisherUniversity of Missouri--Columbiaeng
dc.relation.ispartofcommunityUniversity of Missouri--Columbia. Graduate School. Theses and Dissertationseng
dc.rightsOpenAccess.eng
dc.rights.licenseThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivs 3.0 License.
dc.sourceSubmited to University of Missouri--Columbia Graduate School.eng
dc.titleRelative depth estimation from single monocular images with deep convolutional networkeng
dc.typeThesiseng
thesis.degree.disciplineComputer science (MU)eng
thesis.degree.grantorUniversity of Missouri--Columbiaeng
thesis.degree.levelMasterseng
thesis.degree.nameM.S.eng


Files in this item

[PDF]
[PDF]

This item appears in the following Collection(s)

[-] Show simple item record