WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild
- URL: http://arxiv.org/abs/2409.03753v2
- Date: Mon, 9 Sep 2024 10:04:00 GMT
- Title: WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild
- Authors: Yuntian Deng, Wenting Zhao, Jack Hessel, Xiang Ren, Claire Cardie, Yejin Choi,
- Abstract summary: We introduce WildVis, an interactive tool that enables fast, versatile, and large-scale conversation analysis.
WildVis provides search and visualization capabilities in the text and embedding spaces based on a list of criteria.
We demonstrate WildVis' utility through three case studies: facilitating misuse research, visualizing and comparing topic distributions across datasets, and characterizing user-specific conversation patterns.
- Score: 88.05964311416717
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The increasing availability of real-world conversation data offers exciting opportunities for researchers to study user-chatbot interactions. However, the sheer volume of this data makes manually examining individual conversations impractical. To overcome this challenge, we introduce WildVis, an interactive tool that enables fast, versatile, and large-scale conversation analysis. WildVis provides search and visualization capabilities in the text and embedding spaces based on a list of criteria. To manage million-scale datasets, we implemented optimizations including search index construction, embedding precomputation and compression, and caching to ensure responsive user interactions within seconds. We demonstrate WildVis' utility through three case studies: facilitating chatbot misuse research, visualizing and comparing topic distributions across datasets, and characterizing user-specific conversation patterns. WildVis is open-source and designed to be extendable, supporting additional datasets and customized search and visualization functionalities.
Related papers
- ChatSearch: a Dataset and a Generative Retrieval Model for General Conversational Image Retrieval [31.663016521987764]
We investigate the task of general conversational image retrieval on open-domain images.
To advance this task, we curate a dataset called ChatSearch.
This dataset includes a multi-round multimodal conversational context query for each target image.
We propose a generative retrieval model named ChatSearcher, which is trained end-to-end to accept/produce interleaved image-text inputs/outputs.
arXiv Detail & Related papers (2024-10-24T13:19:22Z) - Automated Question Generation on Tabular Data for Conversational Data Exploration [1.2574534342156884]
We propose a system that recommends interesting questions in natural language based on relevant slices of a dataset in a conversational setting.
We use our own fine-tuned variation of a pre-trained language model(T5) to generate natural language questions in a specific manner.
arXiv Detail & Related papers (2024-07-10T08:07:05Z) - VERA: Generating Visual Explanations of Two-Dimensional Embeddings via Region Annotation [0.0]
Visual Explanations via Region (VERA) is an automatic embedding-annotation approach that generates visual explanations for any two-dimensional embedding.
VERA produces informative explanations that characterize distinct regions in the embedding space, allowing users to gain an overview of the embedding landscape at a glance.
We illustrate the usage of VERA on a real-world data set and validate the utility of our approach with a comparative user study.
arXiv Detail & Related papers (2024-06-07T10:23:03Z) - Open-Vocabulary Camouflaged Object Segmentation [66.94945066779988]
We introduce a new task, open-vocabulary camouflaged object segmentation (OVCOS)
We construct a large-scale complex scene dataset (textbfOVCamo) containing 11,483 hand-selected images with fine annotations and corresponding object classes.
By integrating the guidance of class semantic knowledge and the supplement of visual structure cues from the edge and depth information, the proposed method can efficiently capture camouflaged objects.
arXiv Detail & Related papers (2023-11-19T06:00:39Z) - InViG: Benchmarking Interactive Visual Grounding with 500K Human-Robot
Interactions [23.296139146133573]
We present a large-scale dataset, invig, for interactive visual grounding under language ambiguity.
Our dataset comprises over 520K images accompanied by open-ended goal-oriented disambiguation dialogues.
To the best of our knowledge, the invig dataset is the first large-scale dataset for resolving open-ended interactive visual grounding.
arXiv Detail & Related papers (2023-10-18T17:57:05Z) - V3Det: Vast Vocabulary Visual Detection Dataset [69.50942928928052]
V3Det is a vast vocabulary visual detection dataset with precisely annotated bounding boxes on massive images.
By offering a vast exploration space, V3Det enables extensive benchmarks on both vast and open vocabulary object detection.
arXiv Detail & Related papers (2023-04-07T17:45:35Z) - PRESTO: A Multilingual Dataset for Parsing Realistic Task-Oriented
Dialogs [39.58414649004708]
PRESTO is a dataset of over 550K contextual multilingual conversations between humans and virtual assistants.
It contains challenges that occur in real-world NLU tasks such as disfluencies, code-switching, and revisions.
Our mT5 model based baselines demonstrate that the conversational phenomenon present in PRESTO are challenging to model.
arXiv Detail & Related papers (2023-03-15T21:51:13Z) - Using Large Language Models to Generate Engaging Captions for Data
Visualizations [51.98253121636079]
Large language models (LLM) use sophisticated deep learning technology to produce human-like prose.
Key challenge lies in designing the most effective prompt for the LLM, a task called prompt engineering.
We report on first experiments using the popular LLM GPT-3 and deliver some promising results.
arXiv Detail & Related papers (2022-12-27T23:56:57Z) - OVIS: Open-Vocabulary Visual Instance Search via Visual-Semantic Aligned
Representation Learning [79.49199857462087]
We introduce the task of open-vocabulary visual instance search (OVIS)
Given an arbitrary textual search query, OVIS aims to return a ranked list of visual instances.
We propose to address such a search challenge via visual-semantic aligned representation learning (ViSA)
arXiv Detail & Related papers (2021-08-08T18:13:53Z) - Mobile App Tasks with Iterative Feedback (MoTIF): Addressing Task
Feasibility in Interactive Visual Environments [54.405920619915655]
We introduce Mobile app Tasks with Iterative Feedback (MoTIF), a dataset with natural language commands for the greatest number of interactive environments to date.
MoTIF is the first to contain natural language requests for interactive environments that are not satisfiable.
We perform initial feasibility classification experiments and only reach an F1 score of 37.3, verifying the need for richer vision-language representations.
arXiv Detail & Related papers (2021-04-17T14:48:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.