2018
Authors
Varajão, JE; Cruz Cunha, MM; Martinho, R; Rijo, R; Domingos, D; Peres, E;
Publication
Procedia Computer Science
Abstract
2019
Authors
Marques, P; Padua, L; Adao, T; Hruska, J; Peres, E; Sousa, A; Sousa, JJ;
Publication
REMOTE SENSING
Abstract
Unmanned aerial vehicles have become a popular remote sensing platform for agricultural applications, with an emphasis on crop monitoring. Although there are several methods to detect vegetation through aerial imagery, these remain dependent of manual extraction of vegetation parameters. This article presents an automatic method that allows for individual tree detection and multi-temporal analysis, which is crucial in the detection of missing and new trees and monitoring their health conditions over time. The proposed method is based on the computation of vegetation indices (VIs), while using visible (RGB) and near-infrared (NIR) domain combination bands combined with the canopy height model. An overall segmentation accuracy above 95% was reached, even when RGB-based VIs were used. The proposed method is divided in three major steps: (1) segmentation and first clustering; (2) cluster isolation; and (3) feature extraction. This approach was applied to several chestnut plantations and some parameterssuch as the number of trees present in a plantation (accuracy above 97%), the canopy coverage (93% to 99% accuracy), the tree height (RMSE of 0.33 m and R-2 = 0.86), and the crown diameter (RMSE of 0.44 m and R-2 = 0.96)were automatically extracted. Therefore, by enabling the substitution of time-consuming and costly field campaigns, the proposed method represents a good contribution in managing chestnut plantations in a quicker and more sustainable way.
2019
Authors
Vaz, C; Borges, J; Peres, E; Sousa, J; Reis, MJCS;
Publication
13TH INTERNATIONAL TECHNOLOGY, EDUCATION AND DEVELOPMENT CONFERENCE (INTED2019)
Abstract
E-learning at the University of Tras-os-Montes e Alto Douro (UTAD) is at a very early stage, with the total number of courses available being very low. This is due to the lack of teacher training, lack of technical support, lack of specific tools and lack of institutional support policies, among others. An exception to this is the existence of a platform for teaching support, named SIDE - side.utad.pt, developed at UTAD and which has been systematically used over the last decade. However, SIDE works more like a teaching management platform (e.g. student management, assessment schedules management, attendance management, schedules management, etc.) rather than as a platform for dissemination and management of online content and courses. Recently, we have elsewhere presented a model of a system that is capable of responding to the training needs identified at UTAD, and we have been using it at UTAD. Here, we present the perspectives of the students, using a survey with 23 questions based on a Likert scale, with a 5 points interval, with 1 being the lowest value (or dissatisfaction) and 5 being the maximum (or excellence) value, being the intermediate point considered satisfaction. The questions were grouped in 5 classes: platform, teaching, tests/exams, devices/hardware and global appreciation. The 78 validated answers and their results will be presented and discussed. For the first class, platform, it was concluded that most of the students consider the platform easy to use and that it adapts to the contents that are taught in the various curricular units of the different degrees of study, and that they feel comfortable using and working with the platform and that they recommended higher use by their teachers.
2019
Authors
Morais, R; Silva, N; Mendes, J; Adao, T; Padua, L; Lopez Riquelme, J; Pavon Pulido, N; Sousa, JJ; Peres, E;
Publication
COMPUTERS AND ELECTRONICS IN AGRICULTURE
Abstract
Over the last few years, an extensive set of technologies have been systematically included in precision agriculture (PA) and also in precision viticulture (PV) practices, as tools that allow efficient monitoring of nearly any parameter to achieve sustainable crop management practices and to increase both crop yield and quality. However, many technologies and standards are not yet included on those practices. Therefore, potential benefits that may result from putting together agronomic knowledge with electronics and computer technologies are still not fully accomplished. Both emergent and established paradigms, such as the Internet of Everything (IoE), Internet of Things (IoT), cloud and fog computing, together with increasingly cheaper computing technologies - with very low power requirements and a diversity of wireless technologies, available to exchange data with increased efficiency - and intelligent systems, have evolved to a level where it is virtually possible to expeditiously create and deploy any required monitoring solution. Pushed by all of these technological trends and recent developments, data integration has emerged as the layer between crops and knowledge needed to efficiently manage it. In this paper, the mySense environment is presented, aimed to systematize data acquisition procedures to address common PA/PV issues. mySense builds over a 4-layer technological structure: sensor and sensor nodes, crop field and sensor networks, cloud services and support to front-end applications. It makes available a set of free tools based on the Do-It-Yourself (DIY) concept and enables the use of Arduino (R) and Raspberry Pi (RN) low-cost platforms to quickly prototype a complete monitoring application. Field experiments provide compelling evidences that mySense environment represents an important step forward towards Smart Farming, by enabling the use of low-cost, fast deployment, integrated and transparent technologies to increase PA/PV monitoring applications adoption.
2019
Authors
Adao, T; Padua, L; Marques, P; Sousa, JJ; Peres, E; Magalhaes, L;
Publication
COMPUTERS
Abstract
Virtual models' production is of high pertinence in research and business fields such as architecture, archeology, or video games, whose requirements might range between expeditious virtual building generation for extensively populating computer-based synthesized environments and hypothesis testing through digital reconstructions. There are some known approaches to achieve the production/reconstruction of virtual models, namely digital settlements and buildings. Manual modeling requires highly-skilled manpower and a considerable amount of time to achieve the desired digital contents, in a process composed by many stages that are typically repeated over time. Both image-based and range scanning approaches are more suitable for digital preservation of well-conserved structures. However, they usually require trained human resources to prepare field operations and manipulate expensive equipment (e.g., 3D scanners) and advanced software tools (e.g., photogrammetric applications). To tackle the issues presented by previous approaches, a class of cost-effective, efficient, and scarce-data-tolerant techniques/methods, known as procedural modeling, has been developed aiming at the semi- or fully-automatic production of virtual environments composed of hollow buildings exclusively represented by outer facades or traversable buildings with interiors, either for expeditious generation or reconstruction. Despite the many achievements of the existing procedural modeling approaches, the production of virtual buildings with both interiors and exteriors composed by non-rectangular shapes (convex or concave n-gons) at the floor-plan level is still seldomly addressed. Therefore, a methodology (and respective system) capable of semi-automatically producing ontology-based traversable buildings composed of arbitrarily-shaped floor-plans has been proposed and continuously developed, and is under analysis in this paper, along with its contributions towards the accomplishment of other virtual reality (VR) and augmented reality (AR) projects/works oriented to digital applications for cultural heritage. Recent roof production-related enhancements resorting to the well-established straight skeleton approach are also addressed, as well as forthcoming challenges. The aim is to consolidate this procedural modeling methodology as a valuable computer graphics work and discuss its future directions.
2019
Authors
Adao, T; Pinho, TM; Pádua, L; Santos, N; Sousa, A; Sousa, JJ; Peres, E;
Publication
ISPRS ICWG III/IVA GI4DM 2019 - GEOINFORMATION FOR DISASTER MANAGEMENT
Abstract
Today's climatic proneness to extreme conditions together with human activity have been triggering a series of wildfire-related events that put at risk ecosystems, as well as animal and vegetal patrimony, while threatening dwellers nearby rural or urban areas. When intervention teams-firefighters, civil protection, police-acknowledge these events, usually they have already escalated to proportions hardly controllable mainly due wind gusts, fuel-like solo conditions, among other conditions that propitiate fire spreading. Currently, there is a wide range of camera-capable sensing systems that can be complemented with useful location data-for example, unmanned aerial systems (UAS) integrated cameras and IMU/GPS sensors, stationary surveillance systems-and processing components capable of fostering wildfire events detection and monitoring, thus providing accurate and faithful data for decision support. Precisely in what concerns to detection and monitoring, Deep Learning (DL) has been successfully applied to perform tasks involving classification and/or segmentation of objects of interest in several fields, such as Agriculture, Forestry and other similar areas. Usually, for an effective DL application, more specifically, based on imagery, datasets must rely on heavy and burdensome logistics to gather a representative problem formulation. What if putting together a dataset could be supported in customizable virtual environments, representing faithful situations to train machines, as it already occurs for human training in what regards some particular tasks (rescue operations, surgeries, industry assembling, etc.)? This work intends to propose not only a system to produce faithful virtual environments to complement and/or even supplant the need for dataset gathering logistics while eventually dealing with hypothetical proposals considering climate change events, but also to create tools for synthesizing wildfire environments for DL application. It will therefore enable to extend existing fire datasets with new data generated by human interaction and supervision, viable for training a computational entity. To that end, a study is presented to assess at which extent data virtually generated data can contribute to an effective DL system aiming to identify and segment fire, bearing in mind future developments of active monitoring systems to timely detect fire events and hopefully provide decision support systems to operational teams.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.