Detecting Elevated Air Pollution Levels by Monitoring Web Search Queries
Abstract:
Real-time air pollution monitoring is a valuable tool for public health and environmental surveillance. In recent years, there has been a dramatic increase in air pollution forecasting and monitoring research using artificial neural networks (ANNs). Most of the prior work relied on modeling pollutant concentrations collected from ground-based monitors and meteorological data for long-term forecasting of outdoor ozone, oxides of nitrogen, and PM2.5. Given that traditional, highly sophisticated air quality monitors are expensive and are not universally available, these models cannot adequately serve those not living near pollutant monitoring sites. Furthermore, because prior models were built on physical measurement data collected from sensors, they may not be suitable for predicting public health effects experienced from pollution exposure. This study aims to develop and validate models to nowcast the observed pollution levels using Web search data, which is publicly available in near real-time from major search engines. We developed novel machine learning-based models using both traditional supervised classification methods and state-of-the-art deep learning methods to detect elevated air pollution levels at the US city level, by using generally available meteorological data and aggregate Web-based search volume data derived from Google Trends. We validated the performance of these methods by predicting three critical air pollutants (ozone (O3), nitrogen dioxide (NO2), and fine particulate matter (PM2.5)), across ten major U.S. metropolitan statistical areas (MSAs) in 2017 and 2018.
Download:
Cross-modal Memory Fusion Network for Multimodal Sequential Learning with Missing Values
Abstract:
Information in many real-world applications is inherently multi-modal, sequential and characterized by a variety of missing values. Existing imputation methods mainly focus on the recurrent dynamics in one modality while ignoring the complementary property from other modalities. In this paper, we propose a novel method called cross-modal memory fusion network (CMFN) that explicitly learns both modal-specific and cross-modal dynamics for imputing the missing values in multi-modal sequential learning tasks. Experiments on two datasets demonstrate that our method outperforms state-of-the-art methods and show its potential to better impute missing values in complex multimodal datasets.
Download:
Publications
Chen Lin, Joyce C. Ho, and Eugene Agichtein
Cross-modal Memory Fusion Network for Multimodal Sequential Learning with Missing Values
European Conference on Information Retrieval. Springer, Cham, 2021: 312-319.
Chen Lin, Safoora Yousefi, Elvis Kahoro, Payam Karisani, Donghai Liang, Jeremy Sarnat and Eugene Agichtein
Detecting Elevated Air Pollution Levels by Monitoring Web Search Queries
JMIR Form Res. 2022 Oct 25
Office: R000 // l.beaver@lse.ac.uk // +44 (0)20 7955 - 1111
Postal: CEP/LSE // Houghton Street // London WC2A 2AE // United Kingdom