The ability to correctly identify areas damaged by forest wildfires is essential to plan and monitor the restoration process and estimate the environmental damages after such catastrophic events. The wide availability of satellite data, combined with the recent development of machine learning and deep learning methodologies applied to the computer vision field, makes it extremely interesting to apply the aforementioned techniques to the field of automatic burned area detection. One of the main issues in such a context is the limited amount of labeled data, especially in the context of semantic segmentation. In this paper, we introduce a publicly available dataset for the burned area detection problem for semantic segmentation. The dataset contains 73 satellite images of different forests damaged by wildfires across Europe with a resolution of up to 10m per pixel. Data were collected from the Sentinel-2 L2A satellite mission and the target labels were generated from the Copernicus Emergency Management Service (EMS) annotations, with five different severity levels, ranging from undamaged to completely destroyed. Finally, we report the benchmark values obtained by applying a Convolutional Neural Network on the proposed dataset to address the burned area identification problem.

A Dataset for Burned Area Delineation and Severity Estimation from Satellite Imagery / Colomba, Luca; Farasin, Alessandro; Monaco, Simone; Greco, Salvatore; Garza, Paolo; Apiletti, Daniele; Baralis, Elena; Cerquitelli, Tania. - ELETTRONICO. - (2022), pp. 3893-3897. ((Intervento presentato al convegno International Conference on Information and Knowledge Management (CIKM) 2022 tenutosi a Atlanta (Georgia, USA) nel 17/10/2022 - 21/10/2022 [10.1145/3511808.3557528].

A Dataset for Burned Area Delineation and Severity Estimation from Satellite Imagery

Colomba,Luca;Farasin,Alessandro;Monaco,Simone;Greco,Salvatore;Garza,Paolo;Apiletti,Daniele;Baralis,Elena;Cerquitelli,Tania
2022

Abstract

The ability to correctly identify areas damaged by forest wildfires is essential to plan and monitor the restoration process and estimate the environmental damages after such catastrophic events. The wide availability of satellite data, combined with the recent development of machine learning and deep learning methodologies applied to the computer vision field, makes it extremely interesting to apply the aforementioned techniques to the field of automatic burned area detection. One of the main issues in such a context is the limited amount of labeled data, especially in the context of semantic segmentation. In this paper, we introduce a publicly available dataset for the burned area detection problem for semantic segmentation. The dataset contains 73 satellite images of different forests damaged by wildfires across Europe with a resolution of up to 10m per pixel. Data were collected from the Sentinel-2 L2A satellite mission and the target labels were generated from the Copernicus Emergency Management Service (EMS) annotations, with five different severity levels, ranging from undamaged to completely destroyed. Finally, we report the benchmark values obtained by applying a Convolutional Neural Network on the proposed dataset to address the burned area identification problem.
978-1-4503-9236-5
File in questo prodotto:
File Dimensione Formato  
3511808.3557528.pdf

non disponibili

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 5.69 MB
Formato Adobe PDF
5.69 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
BurnedAreaDatasetPaper.pdf

accesso aperto

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: PUBBLICO - Tutti i diritti riservati
Dimensione 5.15 MB
Formato Adobe PDF
5.15 MB Adobe PDF Visualizza/Apri
Pubblicazioni consigliate

Caricamento pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2971026