Tribe or Not? Critical Inspection of Group Differences Using TribalGram
- URL: http://arxiv.org/abs/2303.09664v1
- Date: Thu, 16 Mar 2023 21:47:48 GMT
- Title: Tribe or Not? Critical Inspection of Group Differences Using TribalGram
- Authors: Yongsu Ahn, Muheng Yan, Yu-Ru Lin, Wen-Ting Chung, Rebecca Hwa
- Abstract summary: Group profiling and group-level analysis have been increasingly used in many domains including policy making and direct marketing.
In this work, we identify a set of accountable group analytics design guidelines to explicate the needs for group differentiation and preventing overgeneralization of a group.
Following the design guidelines, we develop TribalGram, a visual analytic suite that leverages interpretable machine learning algorithms and visualization.
- Score: 13.670327096294479
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rise of AI and data mining techniques, group profiling and
group-level analysis have been increasingly used in many domains including
policy making and direct marketing. In some cases, the statistics extracted
from data may provide insights to a group's shared characteristics; in others,
the group-level analysis can lead to problems including stereotyping and
systematic oppression. How can analytic tools facilitate a more conscientious
process in group analysis? In this work, we identify a set of accountable group
analytics design guidelines to explicate the needs for group differentiation
and preventing overgeneralization of a group. Following the design guidelines,
we develop TribalGram, a visual analytic suite that leverages interpretable
machine learning algorithms and visualization to offer inference assessment,
model explanation, data corroboration, and sense-making. Through the interviews
with domain experts, we showcase how our design and tools can bring a richer
understanding of "groups" mined from the data.
Related papers
- "Show Me What's Wrong!": Combining Charts and Text to Guide Data Analysis [4.016592757754338]
In the context of financial fraud detection, analysts must quickly identify suspicious activity among transactional data.
This is an iterative process made of complex exploratory tasks such as recognizing patterns, grouping, and comparing.
To mitigate the information overload inherent to these steps, we present a tool combining automated information highlights, Large Language Model generated textual insights, and visual analytics.
arXiv Detail & Related papers (2024-10-01T14:16:10Z) - A structured regression approach for evaluating model performance across intersectional subgroups [53.91682617836498]
Disaggregated evaluation is a central task in AI fairness assessment, where the goal is to measure an AI system's performance across different subgroups.
We introduce a structured regression approach to disaggregated evaluation that we demonstrate can yield reliable system performance estimates even for very small subgroups.
arXiv Detail & Related papers (2024-01-26T14:21:45Z) - Outlier-Robust Group Inference via Gradient Space Clustering [50.87474101594732]
Existing methods can improve the worst-group performance, but they require group annotations, which are often expensive and sometimes infeasible to obtain.
We address the problem of learning group annotations in the presence of outliers by clustering the data in the space of gradients of the model parameters.
We show that data in the gradient space has a simpler structure while preserving information about minority groups and outliers, making it suitable for standard clustering methods like DBSCAN.
arXiv Detail & Related papers (2022-10-13T06:04:43Z) - Self-Supervised Visual Representation Learning with Semantic Grouping [50.14703605659837]
We tackle the problem of learning visual representations from unlabeled scene-centric data.
We propose contrastive learning from data-driven semantic slots, namely SlotCon, for joint semantic grouping and representation learning.
arXiv Detail & Related papers (2022-05-30T17:50:59Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - Towards Group Robustness in the presence of Partial Group Labels [61.33713547766866]
spurious correlations between input samples and the target labels wrongly direct the neural network predictions.
We propose an algorithm that optimize for the worst-off group assignments from a constraint set.
We show improvements in the minority group's performance while preserving overall aggregate accuracy across groups.
arXiv Detail & Related papers (2022-01-10T22:04:48Z) - DoGR: Disaggregated Gaussian Regression for Reproducible Analysis of
Heterogeneous Data [4.720638420461489]
We introduce DoGR, a method that discovers latent confounders by simultaneously partitioning the data into overlapping clusters (disaggregation) and modeling the behavior within them (regression)
When applied to real-world data, our method discovers meaningful clusters and their characteristic behaviors, thus giving insight into group differences and their impact on the outcome of interest.
By accounting for latent confounders, our framework facilitates exploratory analysis of noisy, heterogeneous data and can be used to learn predictive models that better generalize to new data.
arXiv Detail & Related papers (2021-08-31T01:58:23Z) - Representation Matters: Assessing the Importance of Subgroup Allocations
in Training Data [85.43008636875345]
We show that diverse representation in training data is key to increasing subgroup performances and achieving population level objectives.
Our analysis and experiments describe how dataset compositions influence performance and provide constructive results for using trends in existing data, alongside domain knowledge, to help guide intentional, objective-aware dataset design.
arXiv Detail & Related papers (2021-03-05T00:27:08Z) - EXPLAIN-IT: Towards Explainable AI for Unsupervised Network Traffic
Analysis [7.447122949368314]
We introduce EXPLAIN-IT, a methodology which deals with unlabeled data, creates meaningful clusters, and suggests an explanation to the clustering results for the end-user.
We apply EXPLAIN-IT to the problem of YouTube video quality classification under encrypted traffic scenarios, showing promising results.
arXiv Detail & Related papers (2020-03-03T17:54:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.