The study combined physical and numerical models, working with data in different formats via a multimodal architecture (photo: Tânia Rego/Agência Brasil)
Developed by researchers at the University of São Paulo in Brazil, the system used the port city of Santos as a sample space, and could enhance the efficiency of civil defense activities in the context of extreme weather events.
Developed by researchers at the University of São Paulo in Brazil, the system used the port city of Santos as a sample space, and could enhance the efficiency of civil defense activities in the context of extreme weather events.
The study combined physical and numerical models, working with data in different formats via a multimodal architecture (photo: Tânia Rego/Agência Brasil)
By José Tadeu Arantes | Agência FAPESP – Predicting extreme events is essential to the preparation and protection of vulnerable regions, especially at a time of climate change. The city of Santos on the coast of São Paulo state (Brazil) is Latin America’s largest port and has been the focus for significant case studies, not least because of the storm surges that threaten its infrastructure and the local ecosystems.
An article reporting the results of a study that focused on a critical part of Santos and used advanced machine learning tools to optimize existing extreme event prediction systems has recently been published in Proceedings of the AAAI Conference on Artificial Intelligence. It mobilized a large number of researchers and was coordinated by Anna Helena Reali Costa, full professor at the University of São Paulo’s Engineering School (POLI-USP). The first author is Marcel Barros, a researcher in POLI-USP’s Department of Computer Engineering and Digital Systems.
The models used to predict sea surface heights, high tides, wave heights and so on are based on differential equations comprising temporal and spatial information such as astronomic tide (determined by the relative positions of the Sun, Moon and Earth), wind regime, current velocity and salinity, among many others.
These models are successful in several areas but they are complex and depend on a number of simplifications and hypotheses. Moreover, new measurements and other data sources cannot always be integrated into them to make forecasts more reliable.
Although modelers are increasingly using machine learning methods capable of identifying patterns in data and extrapolating to new situations, a great many examples are required to train the algorithms that perform complex tasks such as those involved in weather forecasting and storm tide prediction.
“Our study combined the two worlds to develop a model based on machine learning that uses physical models as a starting point but refines them by adding measured data. This research field is known as physics-informed machine learning, or PIML,” Barros explained.
Harmonization of these two sources of information is fundamental to develop more precise and accurate forecasts. However, the use of sensor data faces significant technical challenges, owing especially to its irregular nature and problems such as missing data, temporal displacements, and variations in sampling frequencies. Sensors that fail can take days to be brought back online, but the mechanisms for predicting storm tides must be capable of operating continuously without the missing data.
“To address situations with highly irregular data, we developed an innovative technique to represent the passing of time in neural networks. This representation lets the model be told the position and size of the missing data windows, so that it considers them in its predictions of tide and wave heights,” Barros said.
The innovation permits better modeling of complex natural phenomena and can also be used to model other phenomena that involve irregular time series, such as health data, sensor networks in manufacturing, or financial indicators.
“Furthermore, our model combines different kinds of neural networks so as to integrate multimodal data, such as satellite images, tables and forecasts from numerical models, with possible future integration of other types of data, such as text and audio. This approach is an important step toward more robust and adaptable forecasting systems that can handle the complexity and variability of the data associated with extreme weather events,” Reali Costa said.
The model has three key virtues, she added: it combines physical and numerical models; it represents time in neural networks in a new way; and it works with data in different formats by means of multimodal architecture. “The study offers a methodology that can improve the accuracy of predictions of extreme events, such as storm tides in Santos. At the same time, it highlights the challenges and potential solutions for the integration of physical models and sensor data in complex contexts,” she said.
The study was supported by FAPESP via the Center for Artificial Intelligence (C4AI), an Engineering Research Center (ERC) set up by FAPESP and IBM Brazil, and hosted by POLI-USP.
The article “Early detection of extreme storm tide events using multimodal data processing” is at: ojs.aaai.org/index.php/AAAI/article/view/30194.
The Agency FAPESP licenses news via Creative Commons (CC-BY-NC-ND) so that they can be republished free of charge and in a simple way by other digital or printed vehicles. Agência FAPESP must be credited as the source of the content being republished and the name of the reporter (if any) must be attributed. Using the HMTL button below allows compliance with these rules, detailed in Digital Republishing Policy FAPESP.