Quantifying Spatial Under-reporting Disparities in Resident
Crowdsourcing
- URL: http://arxiv.org/abs/2204.08620v4
- Date: Wed, 6 Dec 2023 00:49:28 GMT
- Title: Quantifying Spatial Under-reporting Disparities in Resident
Crowdsourcing
- Authors: Zhi Liu, Uma Bhandaram, Nikhil Garg
- Abstract summary: We develop a method to identify reporting delays without using external ground-truth data.
We apply our method to over 100,000 resident reports made in New York City and to over 900,000 reports made in Chicago.
- Score: 5.701305404173138
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern city governance relies heavily on crowdsourcing to identify problems
such as downed trees and power lines. A major concern is that residents do not
report problems at the same rates, with heterogeneous reporting delays directly
translating to downstream disparities in how quickly incidents can be
addressed. Here we develop a method to identify reporting delays without using
external ground-truth data. Our insight is that the rates at which duplicate
reports are made about the same incident can be leveraged to disambiguate
whether an incident has occurred by investigating its reporting rate once it
has occurred. We apply our method to over 100,000 resident reports made in New
York City and to over 900,000 reports made in Chicago, finding that there are
substantial spatial and socioeconomic disparities in how quickly incidents are
reported. We further validate our methods using external data and demonstrate
how estimating reporting delays leads to practical insights and interventions
for a more equitable, efficient government service.
Related papers
- How to Train Your Fact Verifier: Knowledge Transfer with Multimodal Open Models [95.44559524735308]
Large language or multimodal model based verification has been proposed to scale up online policing mechanisms for mitigating spread of false and harmful content.
We test the limits of improving foundation model performance without continual updating through an initial study of knowledge transfer.
Our results on two recent multi-modal fact-checking benchmarks, Mocheg and Fakeddit, indicate that knowledge transfer strategies can improve Fakeddit performance over the state-of-the-art by up to 1.7% and Mocheg performance by up to 2.9%.
arXiv Detail & Related papers (2024-06-29T08:39:07Z) - The Impact of Differential Feature Under-reporting on Algorithmic Fairness [86.275300739926]
We present an analytically tractable model of differential feature under-reporting.
We then use to characterize the impact of this kind of data bias on algorithmic fairness.
Our results show that, in real world data settings, under-reporting typically leads to increasing disparities.
arXiv Detail & Related papers (2024-01-16T19:16:22Z) - A Bayesian Spatial Model to Correct Under-Reporting in Urban
Crowdsourcing [1.850972250657274]
Decision-makers often observe the occurrence of events through a reporting process.
We show how to overcome this challenge by leveraging the fact that events are spatially correlated.
arXiv Detail & Related papers (2023-12-18T23:40:56Z) - Evaluation of Crowdsourced Data on Unplowed Roads [0.0]
This study evaluates a novel unplowed roads dataset from the largest crowdsourced transportation data provider Waze.
81% of reports were near known snow events, with false positives occurring at a regular rate of approximately 10 per day statewide.
An effort to encourage unplowed road reporting in Waze through targeted messages on social media did not increase participation.
arXiv Detail & Related papers (2023-10-25T15:48:50Z) - Practitioner-Centric Approach for Early Incident Detection Using
Crowdsourced Data for Emergency Services [2.5328886773979375]
Crowdsourcing platforms such as Waze provide an opportunity for early identification of incidents.
detecting incidents from crowdsourced data streams is difficult due to the challenges of noise and uncertainty associated with such data.
This paper presents a novel problem formulation and solution approach for practitioner-centered incident detection using crowdsourced data.
arXiv Detail & Related papers (2021-12-03T16:51:41Z) - Priority prediction of Asian Hornet sighting report using machine
learning methods [0.0]
The Asian giant hornet (Vespa mandarinia) is devastating not only to native bee colonies, but also to local apiculture.
We propose a method to predict the priority of sighting reports based on machine learning.
arXiv Detail & Related papers (2021-06-28T07:33:53Z) - Competency Problems: On Finding and Removing Artifacts in Language Data [50.09608320112584]
We argue that for complex language understanding tasks, all simple feature correlations are spurious.
We theoretically analyze the difficulty of creating data for competency problems when human bias is taken into account.
arXiv Detail & Related papers (2021-04-17T21:34:10Z) - Emergency Incident Detection from Crowdsourced Waze Data using Bayesian
Information Fusion [4.039649741925056]
This paper presents a novel method for emergency incident detection using noisy crowdsourced Waze data.
We propose a principled computational framework based on observational theory to model the uncertainty in the reliability of crowd-generated reports.
arXiv Detail & Related papers (2020-11-10T22:45:03Z) - Learning to Communicate and Correct Pose Errors [75.03747122616605]
We study the setting proposed in V2VNet, where nearby self-driving vehicles jointly perform object detection and motion forecasting in a cooperative manner.
We propose a novel neural reasoning framework that learns to communicate, to estimate potential errors, and to reach a consensus about those errors.
arXiv Detail & Related papers (2020-11-10T18:19:40Z) - Showing Your Work Doesn't Always Work [73.63200097493576]
"Show Your Work: Improved Reporting of Experimental Results" advocates for reporting the expected validation effectiveness of the best-tuned model.
We analytically show that their estimator is biased and uses error-prone assumptions.
We derive an unbiased alternative and bolster our claims with empirical evidence from statistical simulation.
arXiv Detail & Related papers (2020-04-28T17:59:01Z) - Leveraging Multi-Source Weak Social Supervision for Early Detection of
Fake News [67.53424807783414]
Social media has greatly enabled people to participate in online activities at an unprecedented rate.
This unrestricted access also exacerbates the spread of misinformation and fake news online which might cause confusion and chaos unless being detected early for its mitigation.
We jointly leverage the limited amount of clean data along with weak signals from social engagements to train deep neural networks in a meta-learning framework to estimate the quality of different weak instances.
Experiments on realworld datasets demonstrate that the proposed framework outperforms state-of-the-art baselines for early detection of fake news without using any user engagements at prediction time.
arXiv Detail & Related papers (2020-04-03T18:26:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.