*
 

iForest - Biogeosciences and Forestry

iForest - Biogeosciences and Forestry
*

FIRE-RES Geo-Catch: a mobile application to support reliable fuel mapping at a pan-European scale

iForest - Biogeosciences and Forestry, Volume 16, Issue 5, Pages 268-273 (2023)
doi: https://doi.org/10.3832/ifor4376-016
Published: Oct 19, 2023 - Copyright © 2023 SISEF

Technical Advances

We present a browser-based App for smartphones that is freely available to end-users for collecting geotagged and oriented photos depicting vegetation biomass and fuel characteristics. Our solution builds on advantages of smartphones, allowing their use as easy sensors to collect data by imaging forest ecosystems. The strength and innovation of the proposed solution is based on the following points: (i) using a low memory footprint App, streaming images and data with as little data-volume and memory as needed; (ii) using JavaScript APIs that can be launched from both a browser or as an installed App, as it applies features such as service workers and Progressive Web App; (iii) storing both image and survey data (geolocation and sensor orientation) internally to the device on an indexed database, and synchronizing the data to a cloud-based server when the smartphone is online and when all other safety tests have been successfully passed. The goal is to achieve properly positioned and oriented photos that can be used as training and testing data for future estimation of the surface fuel types based on automatic segmentation and classification via Machine Learning and Deep Learning.

Mobile Data Collection, Citizen Science, Smartphone Position and Orientation, Wildfire, Image Analysis, Progressive Web Apps

  Introduction 

Forest fires are a complex phenomenon driven by different biological, physical, and human factors, at least in southern Europe, experiencing around 85% of the annual burned area on the continent ([6]). In 2022, the burned area increased to 837.000 ha, with countries like Romania, France, and Portugal joining Spain and Italy as the biggest contributors to the burned areas within the EU27 ([17]). Interactions between extreme climate conditions, vegetation, and human behaviour are the most critical and important factors driving forest fires ([5], [4]) not only in southern Europe, but recently in northern and western Europe. Fire activity is also influenced by other variables, such as topography, stand and forest fuel characteristics, size and compartmentalization of forests, land ownership fragmentation, landscape vegetation, management practices, land use changes, rural exodus, and urban expansion ([21], [3]). These variables are mingled in various combinations and influence wildfire risk. Humans through their activities and behaviour affect mostly when and where fires are ignited and how they eventually propagate. In Europe, human activities accounted for more than 95% of ignitions ([10]). Therefore, if new fire prevention-oriented tools could be created and effectively used, they could potentially contribute to the reduction of people caused ignitions and relieve some of the burden that fire management agencies must carry each fire season (less ignitions, less fatigue, better and timely first response).

Several such initiatives have been developed by the forest services, civil protection agencies, academia, and industry, focusing on the importance of effectively mapping fuel types across extended landscapes. The lack of updated and accurate fuel maps for almost all European countries (with the exception of some southern countries) has serious implications for how fire risk is estimated, as it makes it impossible to conduct fire behaviour simulations or run daily fire hazard predictions using similar methods as those applied in the latest version of NFDRS or Wildfire Analyst which is currently implemented in numerous fire agencies.

A concerted project named LANDFIRE, aimed to determine fuel types at the country scale was started many years ago by the Forest Service of the United States using the vegetation fuel groups developed by Anderson ([1]) and Scott & Burgan ([18]); recently, an adaptation to the European landscapes was carried out by Aragoneses et al. ([2]). However, mapping fuel types at larger scales inherently leads to uncertainty in the results, mostly due to the lack of ground data to train the models. The production of a reliable and fine-scale pan-European fuel map requires a rigorous process for validating spatial outputs produced with remote sensing, geoprocessing, Machine Learning (ML), and Deep Learning (DL) techniques with ground-obtained data (i.e., location, photographs, and fuel characterization). Keane & Dickinson ([9]) and Tinkham et al. ([20]) suggested a photoload sampling technique to support similar surface fuel mapping.

Mapping vegetation characteristics, such as fuel types, canopy base height, or biomass, is a daunting task due to the environmental complexity and the features of interest to be identified and distinguished in each landscape. Vegetation can show very different vertical and horizontal spatial characteristics. Grasslands, shrublands, and forests with short to very tall trees can have a vertical structure with high spatial complexity. The latest advancements in remote sensing, from satellites to drones and portable sensors, have remarkable improved our capability to map vegetation parameters; vegetation height and biomass can be estimated by combining remote sensing products with predicting models that use different categories of sophisticated algorithms based on ML and DL. Nevertheless, it is well known that such predictions may suffer from potential errors which can be quite significant, especially when the training process is fed with inaccurate data, the number of training samples is limited, or the attribute to be predicted is not representative of the real-world variability. A solution to the abovementioned issue is to conduct field inventories using mobile-based surveying approaches to support data collection that are necessary for all three steps of AI classification, i.e., training, testing, and validation.

So far, traditional remote sensing products used for fuel mapping in forested lands have been constrained by sensor inability to penetrate the canopy and identify understory features, thus limiting their capacity to capture fuel loads and the distribution of surface fuels ([8]). This limitation is often solved by generating a set of rules considering the type and state of tree vegetation and/or on-site environmental characteristics ([16]). Regardless of the methodology applied, surface fuel classification is based on general rules or variables obtained through statistical models that, while based on empirical data, tend to aggregate predictions around means, missing the opportunity to account for the variability existing in the real world and requiring expert-based adjustments ([12]). This calls for a specific and broad-scale inventory of the main fuel types, especially forest fuels.

Mobile applications for forestry use ([7], [13], [19]) and for wildfire purposes already exist. A concrete example in wildfire purposes is the WFA Pocket ([14]), which implements the classical fire spread models and has been used by thousands of users worldwide. However, to the best of our knowledge, there are not many mobile applications widely used to collect field data and characterize fuel types. The FIRE-RES project aims to integrate fire management approaches through a user-friendly and intituitive mobile application which can be used by non-experts to collect fuel data. Therefore, the main objective of this work is to develop a mobile application to collect oriented and geolocated images in a quick, simple, and reliable manner, thereby collecting user-acquired ground truth data with consistency across large areas and capturing the full range of European fuel types and landscapes. The collected images will then be interpreted by experts using an online viewer and standard guides to translate the images into fuel types. To make the process more rigorous, each image will be interpreted by more than one expert. This redundancy of interpretations will allow for checks on consensus among experts. If an image has different interpretations, a consensus is reached through further interaction among experts. This process will provide a robust way to validate the pan-European fuel map we are planning to generate during our follow-up research.

  Materials and methods 

Overview: the FIRE-RES Geo-Catch App and viewer

The FIRE-RES Geo-Catch framework consists of an App for mobile phones (FIRE-RES Geo-Catch App, hereafter “App”) and a web-GIS portal (FIRE-RES Geo-Catch Viewer, hereafter “web-GIS” or “viewer”), both available online via common browsers. The software was developed as a lightweight and easy-to-use application for a smartphone. By design, it relies on well-supported application programming interfaces (APIs) to collect and share low-memory-footprint online data. The following sections describe the components of the whole framework, including the App, viewer, and storage repositories.

App user interface

The first icon row, respectively from left to right, displays (i) a background light/dark toggle button, (ii) a link button to a manual page that provides detailed information about the App, and (iii) a log button. The latter opens a log-space with a detailed list of messages that provides useful feedback when the App is tested internally and is improved only among FIRE-RES partners.

The second row has a single large button (with a camera icon) that opens the video-stream connecting to the device camera. The last button provides a list of available projects to be chosen by the user, to which their field collected data will be assigned. This option is mandatory, as each captured image must be connected to a project, thus giving greater credibility and reachability to specific data stored on the server; this is useful for meta-analysis, as system administrators can choose at will from which project they can extract the uploaded information. Another mandatory step that must be taken upon the first usage of the App is user validation, described in the next section.

The interface (Fig. 1) keeps a dark background for battery efficiency. Users can change this option by clicking the first icon on the top left of the interface. The importance of this function is not for aesthetic-related reasons but because the App will be used in the field and the light background is a better choice for some outdoor conditions (e.g., strong sunlight).

Fig. 1 - The main App interface. (Left): upon first access from the browser, the user can display a tutorial in the icon “i”; (Center): the user can install the App and prompted to validate an email address; (Right): the user can display of the icon of the App among other mobile applications on the smartphone.

  Enlarge/Shrink   Download   Full Width  Open in Viewer

User validation

To avoid abuse of the App, a white-list approach is used. Photos are loaded and stored on the server only if the user is validated through a verified email address by simply clicking the provided link. To ensure privacy, the email is not stored anywhere but is only used to validate a user’s unique ID (UID), a 10-character alphanumeric string that is persistently assigned to the user’s phone and browser. The same phone will get different UIDs if accessing the App from different browsers on the same device (e.g., Chrome™, Mozilla Firefox, etc.). The users will also get a secret token that they can use to delete their images stored on the server. This validation approach keeps full privacy for the user while preventing anonymous, unwanted, or unauthorized use of the App, e.g., uploading random images or trying to use the uploading capabilities for intrusion in the server.

Collecting geotagged and oriented images

When the user taps the photo icon, the interface for capturing images is activated (Fig. 2). The interface consists of two square areas over a red background (Fig. 2 - left, area A and area B). One square has the video stream from the camera (Fig. 2 - left, area A), overlaid at its top part with coordinates, accuracy, and horizontal and vertical angles in degrees. The accuracy is directly obtained from the APIs, with a 95% confidence level. When the users are satisfied with the image and have checked the geolocation accuracy, they can tap on the video stream square, triggering a snapshot of the video frame and showing it as an image in the bottom square (Fig. 2 - left, area B), along with its coordinates, accuracy, and orientation values (Fig. 2 - central panel). After the user has taken the photos, the Geo-Catch App will double-check that they are accurately geotagged and will inform the user if the camera is not enabled to read and save the location in images. Up to this point, the image is not yet stored, but this step is necessary for double-checking and evaluating the captured image (photo quality and accuracy). Upon approval, the user can tap the captured image (Fig. 2 - left, area B) to finalize it and finally save on the smartphone.

Fig. 2 - From left to right the three steps for capturing and syncing the geolocated image.

  Enlarge/Shrink   Download   Full Width  Open in Viewer

When the device is online, both photo and geolocation information are uploaded to the Interdepartmental Research Center of Geomatics (CIRGEO) server, which belongs to the University of Padua, Italy. If an internet connection is not available, the data package (image with accompanying information) will be cached and synchronized the next time that the FIRE-RES Geo-Catch App detects an internet connection. Users will not lose cached images if the App is shut down, since images captured while offline are stored until synchronized with the server. The local stored copy will be deleted only when the server has successfully received the uploaded image.

Technical details on the workflow are shown in Fig. 3and described below, with user actions in italics:

Fig. 3 - Framework of the workflow from the field to collected and stored geotagged and oriented images.

  Enlarge/Shrink   Download   Full Width  Open in Viewer

  1. the user in the field opens the App” from the device browser or directly from the installed App icon, and this opens the video stream interface (Fig. 3A).
  2. The user tap the video stream box” (Fig. 2 - left, square A). This action triggers the following: (2.1) the image is temporarily saved in the device memory as a JPEG compressed format with 90% compression and shown in the bottom square (Fig. 2 - middle, square B); (2.2) the following additional information is stored as an ASCII string in the EXIF tag “ImageDescription” directly in the JPEG image memory blob: (a) project name; (b) geolocation: longitude and latitude; (c) orientation: horizontal and vertical Euler angles; (d) timestamp in POSIX time: milliseconds from 1° January 1970; (e) horizontal spatial accuracy: in meters; (f) user ID. An example of the information ASCII string below: FIRELinks|9.2174032|44.3598822|358|86|1679734943295|3.2|q7p1w0a1nk.
  3. The user tap the captured image” (Fig. 2 - middle, square B). The tagged image is stored internally on the smartphone using the IndexedDB JavaScript API. The unique key used in the database is the timestamp (refer to point 2.2d above), the value is the image blob, and the information is the ASCII string presented above.
  4. One of two possible automatic actions follows (Fig. 3B). (a) The device is online, which triggers the uploading of the image to the server. If the upload is successful, then the dataset is deleted from the IndexedDB of the device. (b) The device is offline: nothing happens; the image is kept in the internal device database via the IndexedDB JavaScript API.

If event (4b) is true, when the device becomes online later, (4a) is activated. This action is referred to as a “lazy” strategy of synchronizing data. In other words, the action is triggered without the user having to wait or check for the proper upload. If no upload occurs due to the smartphone being offline, the data is safely stored on the user’s device.

Framework

The proposed App can also work completely offline, thanks to its Progressive Web App (PWA) framework. The PWA framework enables web-Apps to work as smartphone Apps without the requirement for the standard Android or iOS frameworks to be installed (i.e., Google Play™, Apple Store™). This has the advantage of working as a simple web-App, thus accessible via a web browser, or as a standalone application.

The App’s activities and interface (Fig. 3B) are built on HTML-JavaScript as the client-side environment and PHP as the server-side environment. The latter collects the raw image data as a byte-array blob and decodes it for storing the image in a file-based structure that is stored on the server. A specific web-GIS portal, FIRE-RES Geo-Catch Viewer, processes all information stored on the server (Fig. 4), thus decoding the EXIF from all images, tagging users to each image, and allowing users to view each photo with its associated data. A potential advantage of this approach is that every image keeps all information, making it completely transferable to other software that might want to process the image using information in the EXIF tag “ImageDescription”.

Fig. 4 - Geo-Catch Viewer is the online portal to access via web-GIS the stored geotagged and oriented images. A current overview of the data collected through the living labs of the FIRE-RES project.

  Enlarge/Shrink   Download   Full Width  Open in Viewer

From photos to fuel types

Our purpose and follow-up plan include the assignment on the acquired photos by experts of a fuel model type that will subsequently be used to train, test, and validate an AI algorithm. This pattern will be provided to a wide range of forestry and wildfire research end-user in the scientific community, enabling them to predict the fuel model from the collected images automatically. Labels will be given by experts who can interpret and assign the correct fuel model type. In order to achieve an objective evaluation, each image will be assessed by two or more different experts, thus cross-validating each interpretation. If all independent experts assign the same fuel model type, then the agreed type will be selected directly; otherwise, a consensus will be reached through interaction among experts. These actions can be quickly deployed via the FIRE-RES Geo-Catch Viewer, which provides the necessary tools for experts to view and label the images directly online and then interact to provide consensus when necessary. Labelled images can be used for training and validation of an AI model that will extract vegetation features from the scene’s background and further classify them into fuel model types. Fig. 5shows the first application of an automatic segmentation process using the method proposed by Kirillov et al. ([11]) in different main fuel types.

Fig. 5 - Example of a segmented phone image using the method of Kirillov et al. ([11]) through main fuel types (A - Forest, B - Shrubland, and C - Grassland).

  Enlarge/Shrink   Download   Full Width  Open in Viewer

Location and orientation accuracy

Recommended image accuracy in terms of geo-position and orientation is related to the targeted scale on several projects. In the FIRE-RES project, we target to validate a map where the cell spatial resolution is 100 × 100 m and the horizontal geolocation accuracy is 10 m (90th percentile Circular Error - CE90), i.e., a minimum of 90 percent of horizontal errors fall within the stated CE90 value. Smartphone’s geolocation accuracy, even under forest canopies, is far below this tolerance threshold, and therefore correction of gross errors is required. It must be noted that this implies filtering out gross errors (outliers). Testing and validation of smartphone expected accuracies have already been conducted in previous research in an urban forest condition ([15]), finding CE95 < 5 m.

Regarding orientation accuracy, rapid tests have been done by orienting the smartphone perpendicular to regular building facades and comparing the direction of horizontal angles using existing cartography. Results showed that the readings fell within ± 20°, but for some of them, an offset of 180° was noticed. This is probably due to the smartphone’s internal interpretation of its orientation regarding the axis of the internal compass. However, additional studies are planned to further investigate this matter.

  Conclusions 

A reliable fuel map is a useful dataset to support the assessment of fire effects and mitigate the negative consequences of wildfires through early warning systems, fire risk estimation, and monitoring in real time the state of forests susceptibility to wildfires. Fuel map classifications are well structured hierarchically and well tested for fire simulations. The FIRE-RES Geo Catch mobile application can contribute to provide more robust and accurate fuel maps at a pan-European scale, improving the accuracy of fuel type classification by using obtained field inventory data. People involved in acquiring geolocated imagery to support fuel map through ground control points will also contribute during validation phases by providing specialized knowledge. The geolocated images need to be well distributed across all over Europe, and our target is to obtain them in a collaborative manner involving forest services, civil protection agencies, research centres, academia, and private companies. Involving the public in general could also be an option, thus taking advantage of more regular data acquisitions. In addition, the databank that will be created through this in-situ crowdsourcing of images can be used for other purposes. Applicability will be further increased if the images are taken recurrently for the same areas during different time periods. Finally, this tool can be used to assess forest status by identifying main natural and anthropogenic alterations, capturing changes not only in fuel types but also in the state and condition of a forest, whether phenological or due to natural or human impacts, identifying the main drivers of forest change, that are not necessarily always driven by fire events.

  Acknowledgements 

This mobile application was funded by the European Union’s Horizon 2020 Research and Innovation Programme, by the project entitled “Innovative technologies & socio-ecological-economic solutions for fire resilient territories in Europe - FIRE-RES”, grant agreement no. 101037419. We would like to thank each of our colleagues who anonymously participated during the testing phase of the App, which has been very helpful in improving and making the App stable and operational.

  References

(1)
Anderson HE (1982). Aids to determining fuel models for estimating fire behavior. General Technical Report INT-122, USDA Forest Service, Inter-mountain Forest and Range Experiment Station, Ogden, UT, USA, pp. 22.
Gscholar
(2)
Aragoneses E, García M, Salis M, Ribeiro LM, Chuvieco E (2023). Classification and mapping of European fuels using a hierarchical, multipurpose fuel classification system. Earth System Science Data 15: 1287-1315.
CrossRef | Gscholar
(3)
Chuvieco E, Pettinari ML, Koutsias N, Forkel M, Hantson S, Turco M (2021). Human and climate drivers of global biomass burning variability. Science of the Total Environment 779: 146361.
CrossRef | Gscholar
(4)
Dupuy JL, Fargeon H, Martin-StPaul N, Pimont F, Ruffault J, Guijarro M, Hernando C, Madrigal J, Fernandes P (2020). Climate change impact on future wildfire danger and activity in southern Europe: a review. Annals of Forest Science 77: 35.
CrossRef | Gscholar
(5)
Fernandes PM, Barros AMG, Pinto A, Santos JA (2016). Characteristics and controls of extremely large wildfires in the western Mediterranean basin. Journal of Geophysical Research: Biogeosciences 121: 2141-2157.
CrossRef | Gscholar
(6)
Ganteaume A, Camia A, Jappiot M, San-Miguel-Ayanz J, Long-Fournel M, Lampin C (2013). A review of the main driving factors of forest fire ignition over Europe. Enviroment Management 51: 651-662.
CrossRef | Gscholar
(7)
Gollob C, Ritter T, Kraßnitzer R, Tockner A, Nothdurft A (2021). Measurement of forest inventory parameters with Apple iPad Pro and integrated LiDAR technology. Remote Sensing 13: 3129.
CrossRef | Gscholar
(8)
Keane RE, Burgan R, Van Wagtendonk J (2001). Mapping wildland fuels for fire management across multiple scales: Integrating remote sensing, GIS, and biophysical modelling. International Journal of Wildland Fire 10: 301-319.
CrossRef | Gscholar
(9)
Keane RE, Dickinson LJ (2007). The photoload sampling technique: estimating surface fuel loadings from downward-looking photographs of synthetic fuelbeds. General Technical Report RMRS-GTR-190, USDA Forest Service, Rocky Mountain Research Station, Fort Collins, CO, USA, pp. 44.
Online | Gscholar
(10)
Khabarov N, Krasovskii A, Obersteiner M, Swart R, Dosio A, San-Miguel-Ayanz J, Durrant T, Camia A, Migliavacca M (2016). Forest fires and adaptation options in Europe. Regional Environmental Change 16: 21-30.
CrossRef | Gscholar
(11)
Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, Xiao T, Whitehead S, Berg AC, Lo WY, Dollár P, Girshick R (2023). Segment anything. arXiv preprint.
CrossRef | Gscholar
(12)
Krsnik G, Busquets Olivé E, Piqué Nicolau M, Larrañaga A, Cardil A, García-Gonzalo J, González-Olabarría JR (2020). Regional level data server for fire hazard evaluation and fuel treatments planning. Forests 12: 4124.
CrossRef | Gscholar
(13)
Mokroš M, Mikita T, Singh A, Tomaštík J, Chudá J, Wezyk P, Kuzelka K, Surovy P, Klimánek M, Zieba-kulawik K, Bobrowski R, Liang X (2021). Novel low-cost mobile mapping systems for forest inventories as terrestrial laser scanning alternatives. International Journal of Applied Earth Observations and Geoinformation 104: 102512.
CrossRef | Gscholar
(14)
Monedero S, Ramirez J, Cardil A (2019). Predicting fire spread and behaviour on the fireline. Wildfire analyst pocket: a mobile app for wildland fire prediction. Ecological Modelling 392: 103-107.
CrossRef | Gscholar
(15)
Pirotti F, Piragnolo M, Guarnieri A, Boscaro M, Cavalli R (2022). Analysis of geospatial behaviour of visitors of urban gardens: is positioning via smartphones a valid solution? In: “Geomatics and Geospatial Technologies” ASITA 2021 (Borgogno-Mondino E, Zamperlin P eds). Communications in Computer and Information Science 10 (1007): 978-3.-030-94426-1_26.
CrossRef | Gscholar
(16)
Rollins MG (2009). LANDFIRE: a national consistent vegetation, wildland fire, and fuel assessment. International Journal of Wildland Fire 18: 235-249.
CrossRef | Gscholar
(17)
San Miguel-Ayanz J, Durrant T, Boca R, Maianti P, Libertá G, Oom D, Branco A, De Rigo D, Ferrari D, Roglia E, Scionti N (2023). Advance report on forest fires in Europe, Middle East and North Africa 2022. JRC Technical Report, Joint Research Centre (European Commission), Publications Office of the European Union, Luxembourg. pp. 47.
CrossRef | Gscholar
(18)
Scott JH, Burgan R (2005). Standard fire behavior fuel models: A comprehensive set for use with Rothermel’s surface fire spread model. General Technical Report RMRS-GTR-153, USDA Forest Service, Rocky Mountain Research Station, Fort Collins, CO, USA, pp. 72.
Online | Gscholar
(19)
Tatsumi S, Yamaguchi K, Furuya N (2022). ForestScanner: a mobile application for measuring and mapping trees with LiDAR-equipped iPhone and iPad. Methods in Ecology and Evolution 00: 1-7.
CrossRef | Gscholar
(20)
Tinkham WT, Hoffman CM, Canfield JM, Vakili E, Reich RM (2016). Using the photoload technique with double sampling to improve surface fuel loading estimates. International Journal of Wildland Fire 25: 224-228.
CrossRef | Gscholar
(21)
Xanthopoulos G, Caballero D, Galante M, Alexandrian D, Rigolot E, Marzano R (2006). Forest fuel management in Europe. Proceedings RMRS-P-41, USDA Forest Service, Rocky Mountain Research Station, Fort Collins, CO, USA, pp. 29.
Gscholar

Authors’ Affiliation

(1)
Erico Kutchartt 0000-0002-9134-4591
Francesco Pirotti 0000-0002-4796-6406
Department of Land, Environment, Agriculture and Forestry - TESAF, University of Padova, v. dell’Università 16, 35020 Legnaro, PD (Italy)
(2)
José Ramón González-Olabarria 0000-0002-5040-712x
Antoni Trasobares 0000-0002-8123-9405
Forest Science and Technology Centre of Catalonia - CTFC, Carretera de Sant Llorenç de Morunys, Km 2, 25280 Solsona (Spain)
(3)
José Ramón González-Olabarria 0000-0002-5040-712x
Sergio de-Miguel 0000-0002-9738-0657
Joint Research Unit CTFC - AGROTECNIO - CERCA, Carretera de Sant Llorenç de Morunys, Km 2, 25280 Solsona (Spain)
(4)
Sergio de-Miguel 0000-0002-9738-0657
Department of Agricultural and Forest Sciences and Engineering, University of Lleida, 25198 Lleida (Spain)
(5)
Adrian Cardil 0000-0002-0185-3959
Tecnosylva, Parque Tecnológico de León, 24009 León (Spain)
(6)
Brigite Botequim 0000-0002-6661-190x
CoLAB ForestWISE - Collaborative Laboratory for Integrated Forest & Fire Management, Quinta de Prados, 5001-801 Vila Real (Portugal)
(7)
Brigite Botequim 0000-0002-6661-190x
Forest Research Centre, School of Agriculture, University of Lisbon, Tapada de Ajuda, 1349-017 Lisbon (Portugal)
(8)
Vassil Vassilev
Space Research and Technology Institute, Bulgarian Academy of Sciences, str. “Acad. Georgy Bonchey” bl. 1, 1113 Sofia (Bulgaria)
(9)
Palaiologos Palaiologou 0000-0001-8507-5201
Department of Forestry and Natural Environment Management, Agricultural University of Athens, 36100 Karpenisi (Greece)
(10)
Martino Rogai
Consiglio Nazionale delle Ricerche - CNR, Istituto per la BioEconomia, v. Madonna del Piano 10, 50019 Sesto Fiorentino, FI (Italy)
(11)
Francesco Pirotti 0000-0002-4796-6406
Interdepartmental Research Center of Geomatics - CIRGEO, University of Padova, v. dell’Università 16, 35020 Legnaro, PD (Italy)

Corresponding author

 
Erico Kutchartt
erico.kutchartt@unipd.it

Citation

Kutchartt E, González-Olabarria JR, Trasobares A, de-Miguel S, Cardil A, Botequim B, Vassilev V, Palaiologou P, Rogai M, Pirotti F (2023). FIRE-RES Geo-Catch: a mobile application to support reliable fuel mapping at a pan-European scale. iForest 16: 268-273. - doi: 10.3832/ifor4376-016

Academic Editor

Marco Borghetti

Paper history

Received: May 09, 2023
Accepted: Jul 10, 2023

First online: Oct 19, 2023
Publication Date: Oct 31, 2023
Publication Time: 3.37 months

© SISEF - The Italian Society of Silviculture and Forest Ecology 2023

  Open Access

This article is distributed under the terms of the Creative Commons Attribution-Non Commercial 4.0 International (https://creativecommons.org/licenses/by-nc/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Creative Commons Licence

Breakdown by View Type

(Waiting for server response...)

Article Usage

Total Article Views: 14134
(from publication date up to now)

Breakdown by View Type
HTML Page Views: 11246
Abstract Page Views: 1856
PDF Downloads: 923
Citation/Reference Downloads: 5
XML Downloads: 104

Web Metrics
Days since publication: 399
Overall contacts: 14134
Avg. contacts per week: 247.96

Article citations are based on data periodically collected from the Clarivate Web of Science web site
(last update: Feb 2023)

(No citations were found up to date. Please come back later)


 

Publication Metrics

by Dimensions ©

List of the papers citing this article based on CrossRef Cited-by.

 

iForest Similar Articles

 

This website uses cookies to ensure you get the best experience on our website. More info