Mass media are a new and important source of information for any natural disaster, mass emergency, pandemic, economic or political event, or extreme weather event affecting one or more communities in a country, providing a relatively high temporal and spatial resolution. The major goal of this research is to show how useful and capable social media is for detecting events in places without actual sensors that could immediately identify a natural hazard. For the entire Italian area, many analytical techniques have been used to determine the spatial and chronological distribution of newspaper articles from Google News about floods and in particular on landslide events. A landslide and flood inventory derived from social media was used as a base proxy to correlate rainfall data and impacts of landslides in an attempt to show how social media in combination with other sources can be utilized to assist government authorities with a better knowledge of the landslide hazard of a territory. Such analysis, further, allowed to outline the resilience at regional scale, considering the number of articles published respect to natural event. In the second and third section of this work, a new data mining technique within Twitter and deep learning technique have been applied. Twitter has been used as source of data for many natural events, but it has rarely been applied to the automatic extraction of landslide events. For filling this gap, Twitter has been investigated to detect data about landslide events in Italian-language. Over 13.000 data were extracted within Twitter considering five keywords referring to landslide news. The dataset was classified manually, providing a solid base for applying the deep learning. The main aim is to obtain an automatic classification on the basis of information about natural hazard. The transformer architecture has been chosen for text classification using the method XLM-RoBERTa. In this work the “Bert For Information on Landslide Events” (BEFILE) model was developed and was tested using 3 different data pre-processing approaches. BEFILE applied without pre-processing showed the highest values of accuracy, equal to 96% and AUC of 0.95.

Analysis of community resilience during natural disasters using data mining on massive social networks exchanges / Rachele Franceschini. - (2023).

Analysis of community resilience during natural disasters using data mining on massive social networks exchanges

Rachele Franceschini
2023

Abstract

Mass media are a new and important source of information for any natural disaster, mass emergency, pandemic, economic or political event, or extreme weather event affecting one or more communities in a country, providing a relatively high temporal and spatial resolution. The major goal of this research is to show how useful and capable social media is for detecting events in places without actual sensors that could immediately identify a natural hazard. For the entire Italian area, many analytical techniques have been used to determine the spatial and chronological distribution of newspaper articles from Google News about floods and in particular on landslide events. A landslide and flood inventory derived from social media was used as a base proxy to correlate rainfall data and impacts of landslides in an attempt to show how social media in combination with other sources can be utilized to assist government authorities with a better knowledge of the landslide hazard of a territory. Such analysis, further, allowed to outline the resilience at regional scale, considering the number of articles published respect to natural event. In the second and third section of this work, a new data mining technique within Twitter and deep learning technique have been applied. Twitter has been used as source of data for many natural events, but it has rarely been applied to the automatic extraction of landslide events. For filling this gap, Twitter has been investigated to detect data about landslide events in Italian-language. Over 13.000 data were extracted within Twitter considering five keywords referring to landslide news. The dataset was classified manually, providing a solid base for applying the deep learning. The main aim is to obtain an automatic classification on the basis of information about natural hazard. The transformer architecture has been chosen for text classification using the method XLM-RoBERTa. In this work the “Bert For Information on Landslide Events” (BEFILE) model was developed and was tested using 3 different data pre-processing approaches. BEFILE applied without pre-processing showed the highest values of accuracy, equal to 96% and AUC of 0.95.
2023
Prof. Filippo Catani
Rachele Franceschini
File in questo prodotto:
File Dimensione Formato  
Thesis_PhD_Rachele_Franceschini_XXXV_cycle.pdf

accesso aperto

Descrizione: Tesi di dottorato Rachele Franceschini
Tipologia: Pdf editoriale (Version of record)
Licenza: Open Access
Dimensione 16.75 MB
Formato Adobe PDF
16.75 MB Adobe PDF

I documenti in FLORE sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificatore per citare o creare un link a questa risorsa: https://hdl.handle.net/2158/1306283
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact