Big data-derived tool facilitates closer monitoring of recovery from natural disasters

big data
Credit: CC0 Public Domain

By analyzing peoples’ visitation patterns to essential establishments like pharmacies, religious centers and grocery stores during Hurricane Harvey, researchers at Texas A&M University have developed a framework to assess the recovery of communities after natural disasters in near real time. They said the information gleaned from their analysis would help federal agencies allocate resources equitably among communities ailing from a disaster.

“Neighboring communities can be impacted very differently after a natural catastrophic event,” said Dr. Ali Mostafavi, associate professor in the Zachry Department of Civil and Environmental Engineering and director of the Urban Resilience.AI Lab. “And so, we need to identify which areas can recover faster than others and which areas are impacted more than others so that we can allocate more resources to areas that need them more.”

The researchers have reported their findings in the Journal of The Royal Society Interface.

The metric that is conventionally used to quantify how communities bounce back from nature-caused setbacks is called resilience and is defined as the ability of a community to return to its pre-disaster state. And so, to measure resilience, factors like the accessibility and distribution of resources, connection between residents within a community and the level of community preparedness for an unforeseen disaster are critical.

The standard way of obtaining data needed to estimate resilience is through surveys. The questions considered, among many others, are how and to what extent businesses or households were affected by the natural disaster and the stage of recovery. However, Mostafavi said these survey-based methods, although extremely useful, take a long time to conduct, with the results of the survey becoming available many months after the disaster.

“For federal agencies allocating funds, recovery information is actually needed in a faster and more near real-time fashion for communities that are trailing in the recovery process,” said Mostafavi. “The solution, we thought, was to look for emerging sources of data other than surveys that could provide more granular insights into community recovery at a scale not previously investigated.”

Mostafavi and his collaborators turned to community-level big data, particularly the information collected by companies that keep track of visits to locations within a perimeter from anonymized cell phone data. In particular, the researchers partnered with a company called SafeGraph to obtain location data for the people in Harris County, Texas, around the time of Hurricane Harvey. As a first step, they determined “points of interest” corresponding to the locations of establishments, like hospitals, gas stations and stores, that might experience a change in visitor traffic due to the hurricane.

Next, the researchers mined the big data and obtained the number of visits to each point of interest before and during the hurricane. For different communities in Harris County, they calculated the time taken for the visits to return to the pre-disaster level and the general resilience, that is, the combined resilience of each point of interest based on the percent change in the number of visits due to the hurricane.

Their analysis revealed that communities that had low resilience also experienced more flooding. However, their results also showed that the level of impact did not necessarily correlate with recovery.

“It’s intuitive to assume, for example, that businesses impacted more will have slower recovery, which actually wasn’t the case,” said Mostafavi. “There were places where visits dropped significantly, but they recovered fast. But then others that were impacted less but took longer to recover, which indicated the importance of both time and general resilience in evaluating a community’s recovery.”

The researchers also noted that another important finding was that the areas that are in close proximity to those that had flooding are also impacted, suggesting that the spatial reach of flooding goes beyond flooded areas.

“Although we focused on Hurricane Harvey for this study, our framework is applicable for any other natural disaster as well,” said Mostafavi. “But as a next step, we’d like to create an intelligent dashboard that would display the rate of recovery and impacts in different areas in near real time and also predict the likelihood of future access disruption and recovery patterns after a heavy downpour.”


Civil engineers find link between hospitals and schools key to community resilience


More information:
Cristian Podesta et al, Quantifying community resilience based on fluctuations in visits to points-of-interest derived from digital trace data, Journal of The Royal Society Interface (2021). DOI: 10.1098/rsif.2021.0158

Provided by
Texas A&M University


Citation:
Big data-derived tool facilitates closer monitoring of recovery from natural disasters (2021, July 22)
retrieved 22 July 2021
from https://phys.org/news/2021-07-big-data-derived-tool-closer-recovery.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

For all the latest Science News Click Here 

 For the latest news and updates, follow us on Google News

Read original article here

Denial of responsibility! TheDailyCheck is an automatic aggregator around the global media. All the content are available free on Internet. We have just arranged it in one platform for educational purpose only. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials on our website, please contact us by email – [email protected] The content will be deleted within 24 hours.