Assessing the trade-off between prediction accuracy and interpretability
for topic modeling on energetic materials corpora
- URL: http://arxiv.org/abs/2206.00773v1
- Date: Wed, 1 Jun 2022 21:28:21 GMT
- Title: Assessing the trade-off between prediction accuracy and interpretability
for topic modeling on energetic materials corpora
- Authors: Monica Puerto, Mason Kellett, Rodanthi Nikopoulou, Mark D. Fuge, Ruth
Doherty, Peter W. Chung, and Zois Boukouvalas
- Abstract summary: We study the trade-off between prediction accuracy and interpretability by implementing three document embedding methods.
This study was carried out on a novel labeled energetics dataset created and validated by our team of energetics experts.
- Score: 2.1694433437280765
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the amount and variety of energetics research increases, machine aware
topic identification is necessary to streamline future research pipelines. The
makeup of an automatic topic identification process consists of creating
document representations and performing classification. However, the
implementation of these processes on energetics research imposes new
challenges. Energetics datasets contain many scientific terms that are
necessary to understand the context of a document but may require more complex
document representations. Secondly, the predictions from classification must be
understandable and trusted by the chemists within the pipeline. In this work,
we study the trade-off between prediction accuracy and interpretability by
implementing three document embedding methods that vary in computational
complexity. With our accuracy results, we also introduce local interpretability
model-agnostic explanations (LIME) of each prediction to provide a localized
understanding of each prediction and to validate classifier decisions with our
team of energetics experts. This study was carried out on a novel labeled
energetics dataset created and validated by our team of energetics experts.
Related papers
- Probing the limitations of multimodal language models for chemistry and materials research [3.422786943576035]
We introduce MaCBench, a benchmark for evaluating how vision-language models handle real-world chemistry and materials science tasks.
We find that while these systems show promising capabilities in basic perception tasks, they exhibit fundamental limitations in spatial reasoning, cross-modal information synthesis, and logical inference.
Our insights have important implications beyond chemistry and materials science, suggesting that developing reliable multimodal AI scientific assistants may require advances in curating suitable training data and approaches to training those models.
arXiv Detail & Related papers (2024-11-25T21:51:45Z) - Likelihood as a Performance Gauge for Retrieval-Augmented Generation [78.28197013467157]
We show that likelihoods serve as an effective gauge for language model performance.
We propose two methods that use question likelihood as a gauge for selecting and constructing prompts that lead to better performance.
arXiv Detail & Related papers (2024-11-12T13:14:09Z) - Causal Representation Learning from Multimodal Biological Observations [57.00712157758845]
We aim to develop flexible identification conditions for multimodal data.
We establish identifiability guarantees for each latent component, extending the subspace identification results from prior work.
Our key theoretical ingredient is the structural sparsity of the causal connections among distinct modalities.
arXiv Detail & Related papers (2024-11-10T16:40:27Z) - BLADE: Benchmarking Language Model Agents for Data-Driven Science [18.577658530714505]
LM-based agents equipped with planning, memory, and code execution capabilities have the potential to support data-driven science.
We present BLADE, a benchmark to automatically evaluate agents' multifaceted approaches to open-ended research questions.
arXiv Detail & Related papers (2024-08-19T02:59:35Z) - From Pixels to Insights: A Survey on Automatic Chart Understanding in the Era of Large Foundation Models [98.41645229835493]
Data visualization in the form of charts plays a pivotal role in data analysis, offering critical insights and aiding in informed decision-making.
Large foundation models, such as large language models, have revolutionized various natural language processing tasks.
This survey paper serves as a comprehensive resource for researchers and practitioners in the fields of natural language processing, computer vision, and data analysis.
arXiv Detail & Related papers (2024-03-18T17:57:09Z) - Transformers and Language Models in Form Understanding: A Comprehensive
Review of Scanned Document Analysis [16.86139440201837]
We focus on the topic of form understanding in the context of scanned documents.
Our research methodology involves an in-depth analysis of popular documents and forms of understanding of trends over the last decade.
We showcase how transformers have propelled the field forward, revolutionizing form-understanding techniques.
arXiv Detail & Related papers (2024-03-06T22:22:02Z) - NLP for Knowledge Discovery and Information Extraction from Energetics
Corpora [0.0]
We present a demonstration of the utility of NLP for aiding research into energetic materials and associated systems.
The NLP method enables machine understanding of textual data, offering an automated route to knowledge discovery and information extraction.
arXiv Detail & Related papers (2024-02-10T14:43:08Z) - Generative Judge for Evaluating Alignment [84.09815387884753]
We propose a generative judge with 13B parameters, Auto-J, designed to address these challenges.
Our model is trained on user queries and LLM-generated responses under massive real-world scenarios.
Experimentally, Auto-J outperforms a series of strong competitors, including both open-source and closed-source models.
arXiv Detail & Related papers (2023-10-09T07:27:15Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Data Mining in Clinical Trial Text: Transformers for Classification and
Question Answering Tasks [2.127049691404299]
This research applies advances in natural language processing to evidence synthesis based on medical texts.
The main focus is on information characterized via the Population, Intervention, Comparator, and Outcome (PICO) framework.
Recent neural network architectures based on transformers show capacities for transfer learning and increased performance on downstream natural language processing tasks.
arXiv Detail & Related papers (2020-01-30T11:45:59Z) - ORB: An Open Reading Benchmark for Comprehensive Evaluation of Machine
Reading Comprehension [53.037401638264235]
We present an evaluation server, ORB, that reports performance on seven diverse reading comprehension datasets.
The evaluation server places no restrictions on how models are trained, so it is a suitable test bed for exploring training paradigms and representation learning.
arXiv Detail & Related papers (2019-12-29T07:27:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.