|
Patricia L. Suarez, D. C., Angel Sappa. (2023). Boosting Guided Super-Resolution Performance with Synthesized Images. In 17th International Conference On Signal Image Technology & Internet Based Systems.
|
|
|
Carlos Monsalve, Alain April, & Alain Abran. (2011). BPM and requirements elicitation at multiple levels of abstraction: A review. In IADIS International Conference on Information Systems 2011 (pp. 237–242).
Abstract: Business process models can be useful for requirements elicitation, among other things. Software development depends on the quality of the requirements elicitation activities, and so adequately modeling business processes (BPs) is critical. A key factor in achieving this is the active participation of all the stakeholders in the development of a shared vision of BPs.
Unfortunately, organizations often find themselves left with inconsistent BPs that do not cover all the stakeholders’ needs
and constraints. However, consolidation of the various stakeholder requirements may be facilitated through the use of multiple levels of abstraction (MLA). This article contributes to the research into MLA use in business process modeling (BPM) for software requirements by reviewing the theoretical foundations of MLA and their use in various BP-oriented approaches.
|
|
|
Charco, J. L., Sappa, A.D., Vintimilla, B.X., Velesaca, H.O. (2021). Camera pose estimation in multi-view environments:from virtual scenarios to the real world. In Image and Vision Computing Journal. (Article number 104182), Vol. 110.
Abstract: This paper presents a domain adaptation strategy to efficiently train network architectures for estimating the relative camera pose in multi-view scenarios. The network architectures are fed by a pair of simultaneously acquired
images, hence in order to improve the accuracy of the solutions, and due to the lack of large datasets with pairs of
overlapped images, a domain adaptation strategy is proposed. The domain adaptation strategy consists on transferring the knowledge learned from synthetic images to real-world scenarios. For this, the networks are firstly
trained using pairs of synthetic images, which are captured at the same time by a pair of cameras in a virtual environment; and then, the learned weights of the networks are transferred to the real-world case, where the networks are retrained with a few real images. Different virtual 3D scenarios are generated to evaluate the
relationship between the accuracy on the result and the similarity between virtual and real scenarios—similarity
on both geometry of the objects contained in the scene as well as relative pose between camera and objects in the
scene. Experimental results and comparisons are provided showing that the accuracy of all the evaluated networks for estimating the camera pose improves when the proposed domain adaptation strategy is used,
highlighting the importance on the similarity between virtual-real scenarios.
|
|
|
Patricia L. Suarez, Angel D. Sappa, & Boris X. Vintimilla. (2017). Colorizing Infrared Images through a Triplet Condictional DCGAN Architecture. In 19th International Conference on Image Analysis and Processing. (pp. 287–297).
|
|
|
Armin Mehri, & Angel D. Sappa. (2019). Colorizing Near Infrared Images through a Cyclic Adversarial Approach of Unpaired Samples. In Conference on Computer Vision and Pattern Recognition Workshops (CVPR 2019); Long Beach, California, United States (pp. 971–979).
Abstract: This paper presents a novel approach for colorizing
near infrared (NIR) images. The approach is based on
image-to-image translation using a Cycle-Consistent adversarial network for learning the color channels on unpaired dataset. This architecture is able to handle unpaired datasets. The approach uses as generators tailored
networks that require less computation times, converge
faster and generate high quality samples. The obtained results have been quantitatively—using standard evaluation
metrics—and qualitatively evaluated showing considerable
improvements with respect to the state of the art
|
|
|
Velesaca, H. O., Suárez, P. L., Mira, R., & Sappa, A.D. (2021). Computer Vision based Food Grain Classification: a Comprehensive Survey. In Computers and Electronics in Agriculture Journal. (Article number 106287), Vol. 187.
|
|
|
Roberto Jacome Galarza, Miguel-Andrés Realpe-Robalino, Chamba-Eras LuisAntonio, & Viñán-Ludeña MarlonSantiago and Sinche-Freire Javier-Francisco. (2019). Computer vision for image understanding. A comprehensive review. In International Conference on Advances in Emerging Trends and Technologies (ICAETT 2019); Quito, Ecuador (pp. 248–259).
Abstract: Computer Vision has its own Turing test: Can a machine describe the contents of an image or a video in the way a human being would do? In this paper, the progress of Deep Learning for image recognition is analyzed in order to know the answer to this question. In recent years, Deep Learning has increased considerably the precision rate of many tasks related to computer vision. Many datasets of labeled images are now available online, which leads to pre-trained models for many computer vision applications. In this work, we gather information of the latest techniques to perform image understanding and description. As a conclusion we obtained that the combination of Natural Language Processing (using Recurrent Neural Networks and Long Short-Term Memory) plus Image Understanding (using Convolutional Neural Networks) could bring new types of powerful and useful applications in which the computer will be able to answer questions about the content of images and videos. In order to build datasets of labeled images, we need a lot of work and most of the datasets are built using crowd work. These new applications have the potential to increase the human machine interaction to new levels of usability and user’s satisfaction.
|
|
|
Michael Teutsch, A. S. & R. H. (2021). Computer Vision in the Infrared Spectrum: Challenges and ApproachesComputer Vision in the Infrared Spectrum: Challenges and Approaches. Synthesis Lectures on Computer Vision, Vol. 10 No. 2, pp. 138.
|
|
|
M. Oliveira, L. Seabra Lopes, G. Hyun Lim, S. Hamidreza Kasaei, Angel D. Sappa, & A. Tomé. (2015). Concurrent Learning of Visual Codebooks and Object Categories in Open- ended Domains. In Intelligent Robots and Systems (IROS), 2015 IEEE/RSJ International Conference on, Hamburg, Germany, 2015 (pp. 2488–2495). Hamburg, Germany: IEEE.
Abstract: In open-ended domains, robots must continuously learn new object categories. When the training sets are created offline, it is not possible to ensure their representativeness with respect to the object categories and features the system will find when operating online. In the Bag of Words model, visual codebooks are usually constructed from training sets created offline. This might lead to non-discriminative visual words and, as a consequence, to poor recognition performance. This paper proposes a visual object recognition system which concurrently learns in an incremental and online fashion both the visual object category representations as well as the codebook words used to encode them. The codebook is defined using Gaussian Mixture Models which are updated using new object views. The approach contains similarities with the human visual object recognition system: evidence suggests that the development of recognition capabilities occurs on multiple levels and is sustained over large periods of time. Results show that the proposed system with concurrent learning of object categories and codebooks is capable of learning more categories, requiring less examples, and with similar accuracies, when compared to the classical Bag of Words approach using codebooks constructed offline.
|
|
|
Patricia Súarez, H. V., Dario Carpio & Angel Sappa. (2023). Corn Kernel Classification From Few Training Samples. In journal Artificial Intelligence in Agriculture, Vol. 9, pp. 89–99.
|
|