Home | << 1 2 3 4 5 6 7 >> |
Records | |||||
---|---|---|---|---|---|
Author | Cristhian A. Aguilera, Cristhian Aguilera, Cristóbal A. Navarro, & Angel D. Sappa | ||||
Title | Fast CNN Stereo Depth Estimation through Embedded GPU Devices | Type | Journal Article | ||
Year | 2020 | Publication | Sensors 2020 | Abbreviated Journal | |
Volume | Vol. 2020-June | Issue | 11 | Pages | pp. 1-13 |
Keywords | stereo matching; deep learning; embedded GPU | ||||
Abstract | Current CNN-based stereo depth estimation models can barely run under real-time constraints on embedded graphic processing unit (GPU) devices. Moreover, state-of-the-art evaluations usually do not consider model optimization techniques, being that it is unknown what is the current potential on embedded GPU devices. In this work, we evaluate two state-of-the-art models on three different embedded GPU devices, with and without optimization methods, presenting performance results that illustrate the actual capabilities of embedded GPU devices for stereo depth estimation. More importantly, based on our evaluation, we propose the use of a U-Net like architecture for postprocessing the cost-volume, instead of a typical sequence of 3D convolutions, drastically augmenting the runtime speed of current models. In our experiments, we achieve real-time inference speed, in the range of 5–32 ms, for 1216 368 input stereo images on the Jetson TX2, Jetson Xavier, and Jetson Nano embedded devices. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 14248220 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 132 | ||
Permanent link to this record | |||||
Author | Rafael E. Rivadeneira; Patricia L. Suarez; Angel D. Sappa; Boris X. Vintimilla. | ||||
Title | Thermal Image SuperResolution through Deep Convolutional Neural Network. | Type | Conference Article | ||
Year | 2019 | Publication | 16th International Conference on Image Analysis and Recognition (ICIAR 2019); Waterloo, Canadá | Abbreviated Journal | |
Volume | Issue | Pages | 417-426 | ||
Keywords | |||||
Abstract | Due to the lack of thermal image datasets, a new dataset has been acquired for proposed a superesolution approach using a Deep Convolution Neural Network schema. In order to achieve this image enhancement process a new thermal images dataset is used. Di?erent experiments have been carried out, ?rstly, the proposed architecture has been trained using only images of the visible spectrum, and later it has been trained with images of the thermal spectrum, the results showed that with the network trained with thermal images, better results are obtained in the process of enhancing the images, maintaining the image details and perspective. The thermal dataset is available at http://www.cidis.espol.edu.ec/es/dataset | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | gtsi @ user @ | Serial | 103 | ||
Permanent link to this record | |||||
Author | M. Oliveira; L. Seabra Lopes; G. Hyun Lim; S. Hamidreza Kasaei; Angel D. Sappa; A. Tomé | ||||
Title | Concurrent Learning of Visual Codebooks and Object Categories in Open- ended Domains | Type | Conference Article | ||
Year | 2015 | Publication | Intelligent Robots and Systems (IROS), 2015 IEEE/RSJ International Conference on, Hamburg, Germany, 2015 | Abbreviated Journal | |
Volume | Issue | Pages | 2488 - 2495 | ||
Keywords | Birds, Training, Legged locomotion, Visualization, Histograms, Object recognition, Gaussian mixture model | ||||
Abstract | In open-ended domains, robots must continuously learn new object categories. When the training sets are created offline, it is not possible to ensure their representativeness with respect to the object categories and features the system will find when operating online. In the Bag of Words model, visual codebooks are usually constructed from training sets created offline. This might lead to non-discriminative visual words and, as a consequence, to poor recognition performance. This paper proposes a visual object recognition system which concurrently learns in an incremental and online fashion both the visual object category representations as well as the codebook words used to encode them. The codebook is defined using Gaussian Mixture Models which are updated using new object views. The approach contains similarities with the human visual object recognition system: evidence suggests that the development of recognition capabilities occurs on multiple levels and is sustained over large periods of time. Results show that the proposed system with concurrent learning of object categories and codebooks is capable of learning more categories, requiring less examples, and with similar accuracies, when compared to the classical Bag of Words approach using codebooks constructed offline. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | IEEE | Place of Publication | Hamburg, Germany | Editor | |
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) | ||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 41 | ||
Permanent link to this record | |||||
Author | Miguel Oliveira; Vítor Santos; Angel D. Sappa; Paulo Dias | ||||
Title | Scene representations for autonomous driving: an approach based on polygonal primitives | Type | Conference Article | ||
Year | 2015 | Publication | Iberian Robotics Conference (ROBOT 2015), Lisbon, Portugal, 2015 | Abbreviated Journal | |
Volume | 417 | Issue | Pages | 503-515 | |
Keywords | Scene reconstruction, Point cloud, Autonomous vehicles | ||||
Abstract | In this paper, we present a novel methodology to compute a 3D scene representation. The algorithm uses macro scale polygonal primitives to model the scene. This means that the representation of the scene is given as a list of large scale polygons that describe the geometric structure of the environment. Results show that the approach is capable of producing accurate descriptions of the scene. In addition, the algorithm is very efficient when compared to other techniques. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Springer International Publishing Switzerland 2016 | Place of Publication | Editor | ||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | Second Iberian Robotics Conference | ||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 45 | ||
Permanent link to this record | |||||
Author | Cristhian A. Aguilera; Cristhian Aguilera; Angel D. Sappa | ||||
Title | Melamine faced panels defect classification beyond the visible spectrum. | Type | Journal Article | ||
Year | 2018 | Publication | In Sensors 2018 | Abbreviated Journal | |
Volume | Vol. 11 | Issue | Issue 11 | Pages | |
Keywords | |||||
Abstract | In this work, we explore the use of images from different spectral bands to classify defects in melamine faced panels, which could appear through the production process. Through experimental evaluation, we evaluate the use of images from the visible (VS), near-infrared (NIR), and long wavelength infrared (LWIR), to classify the defects using a feature descriptor learning approach together with a support vector machine classifier. Two descriptors were evaluated, Extended Local Binary Patterns (E-LBP) and SURF using a Bag of Words (BoW) representation. The evaluation was carried on with an image set obtained during this work, which contained five different defect categories that currently occurs in the industry. Results show that using images from beyond the visual spectrum helps to improve classification performance in contrast with a single visible spectrum solution. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | gtsi @ user @ | Serial | 89 | ||
Permanent link to this record | |||||
Author | Xavier Soria; Angel D. Sappa; Riad Hammoud | ||||
Title | Wide-Band Color Imagery Restoration for RGB-NIR Single Sensor Image. Sensors 2018 ,2059. | Type | Journal Article | ||
Year | 2018 | Publication | Abbreviated Journal | ||
Volume | Vol. 18 | Issue | Issue 7 | Pages | |
Keywords | |||||
Abstract | Multi-spectral RGB-NIR sensors have become ubiquitous in recent years. These sensors allow the visible and near-infrared spectral bands of a given scene to be captured at the same time. With such cameras, the acquired imagery has a compromised RGB color representation due to near-infrared bands (700–1100 nm) cross-talking with the visible bands (400–700 nm). This paper proposes two deep learning-based architectures to recover the full RGB color images, thus removing the NIR information from the visible bands. The proposed approaches directly restore the high-resolution RGB image by means of convolutional neural networks. They are evaluated with several outdoor images; both architectures reach a similar performance when evaluated in different scenarios and using different similarity metrics. Both of them improve the state of the art approaches. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | gtsi @ user @ | Serial | 96 | ||
Permanent link to this record | |||||
Author | Angel Morera; Angel Sánchez; Angel D. Sappa; José F. Vélez | ||||
Title | Robust Detection of Outdoor Urban Advertising Panels in Static Images. | Type | Conference Article | ||
Year | 2019 | Publication | 17th International Conference on Practical Applications of Agents and Multi-Agent Systems (PAAMS 2019); Ávila, España. Communications in Computer and Information Science | Abbreviated Journal | |
Volume | 1047 | Issue | Pages | 246-256 | |
Keywords | |||||
Abstract | One interesting publicity application for Smart City environments is recognizing brand information contained in urban advertising panels. For such a purpose, a previous stage is to accurately detect and locate the position of these panels in images. This work presents an effective solution to this problem using a Single Shot Detector (SSD) based on a deep neural network architecture that minimizes the number of false detections under multiple variable conditions regarding the panels and the scene. Achieved experimental results using the Intersection over Union (IoU) accuracy metric make this proposal applicable in real complex urban images. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | gtsi @ user @ | Serial | 107 | ||
Permanent link to this record | |||||
Author | Julien Poujol; Cristhian A. Aguilera; Etienne Danos; Boris X. Vintimilla; Ricardo Toledo; Angel D. Sappa | ||||
Title | A visible-Thermal Fusion based Monocular Visual Odometry | Type | Conference Article | ||
Year | 2015 | Publication | Iberian Robotics Conference (ROBOT 2015), International Conference on, Lisbon, Portugal, 2015 | Abbreviated Journal | |
Volume | 417 | Issue | Pages | 517-528 | |
Keywords | Monocular Visual Odometry; LWIR-RGB cross-spectral Imaging; Image Fusion | ||||
Abstract | The manuscript evaluates the performance of a monocular visual odometry approach when images from different spectra are considered, both independently and fused. The objective behind this evaluation is to analyze if classical approaches can be improved when the given images, which are from different spectra, are fused and represented in new domains. The images in these new domains should have some of the following properties: i) more robust to noisy data; ii) less sensitive to changes (e.g., lighting); iii) more rich in descriptive information, among other. In particular in the current work two different image fusion strategies are considered. Firstly, images from the visible and thermal spectrum are fused using a Discrete Wavelet Transform (DWT) approach. Secondly, a monochrome threshold strategy is considered. The obtained representations are evaluated under a visual odometry framework, highlighting their advantages and disadvantages, using different urban and semi-urban scenarios. Comparisons with both monocular-visible spectrum and monocular-infrared spectrum, are also provided showing the validity of the proposed approach. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 44 | ||
Permanent link to this record | |||||
Author | Cristhian A. Aguilera; Francisco J. Aguilera; Angel D. Sappa; Ricardo Toledo | ||||
Title | Learning crossspectral similarity measures with deep convolutional neural networks | Type | Conference Article | ||
Year | 2016 | Publication | IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) Workshops | Abbreviated Journal | |
Volume | Issue | Pages | 267-275 | ||
Keywords | |||||
Abstract | The simultaneous use of images from different spectra can be helpful to improve the performance of many com- puter vision tasks. The core idea behind the usage of cross- spectral approaches is to take advantage of the strengths of each spectral band providing a richer representation of a scene, which cannot be obtained with just images from one spectral band. In this work we tackle the cross-spectral image similarity problem by using Convolutional Neural Networks (CNNs). We explore three different CNN archi- tectures to compare the similarity of cross-spectral image patches. Specifically, we train each network with images from the visible and the near-infrared spectrum, and then test the result with two public cross-spectral datasets. Ex- perimental results show that CNN approaches outperform the current state-of-art on both cross-spectral datasets. Ad- ditionally, our experiments show that some CNN architec- tures are capable of generalizing between different cross- spectral domains. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 48 | ||
Permanent link to this record | |||||
Author | N. Onkarappa; Cristhian A. Aguilera; B. X. Vintimilla; Angel D. Sappa | ||||
Title | Cross-spectral Stereo Correspondence using Dense Flow Fields | Type | Conference Article | ||
Year | 2014 | Publication | Computer Vision Theory and Applications (VISAPP), 2014 International Conference on, Lisbon, Portugal, 2014 | Abbreviated Journal | |
Volume | 3 | Issue | Pages | 613 - 617 | |
Keywords | Cross-spectral Stereo Correspondence, Dense Optical Flow, Infrared and Visible Spectrum | ||||
Abstract | This manuscript addresses the cross-spectral stereo correspondence problem. It proposes the usage of a dense flow field based representation instead of the original cross-spectral images, which have a low correlation. In this way, working in the flow field space, classical cost functions can be used as similarity measures. Preliminary experimental results on urban environments have been obtained showing the validity of the proposed approach. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | IEEE | Place of Publication | Editor | ||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | 2014 International Conference on Computer Vision Theory and Applications (VISAPP) | ||
Notes | Approved | no | |||
Call Number | cidis @ cidis @ | Serial | 27 | ||
Permanent link to this record |