Unlocking Model Insights: A Dataset for Automated Model Card Generation
- URL: http://arxiv.org/abs/2309.12616v1
- Date: Fri, 22 Sep 2023 04:46:11 GMT
- Title: Unlocking Model Insights: A Dataset for Automated Model Card Generation
- Authors: Shruti Singh, Hitesh Lodwal, Husain Malwat, Rakesh Thakur, Mayank
Singh
- Abstract summary: We introduce a dataset of 500 question-answer pairs for 25 ML models.
We employ annotators to extract the answers from the original paper.
Our experiments with ChatGPT-3.5, LLaMa, and Galactica showcase a significant gap in the understanding of research papers by these LMs.
- Score: 4.167070553534516
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Language models (LMs) are no longer restricted to ML community, and
instruction-tuned LMs have led to a rise in autonomous AI agents. As the
accessibility of LMs grows, it is imperative that an understanding of their
capabilities, intended usage, and development cycle also improves. Model cards
are a popular practice for documenting detailed information about an ML model.
To automate model card generation, we introduce a dataset of 500
question-answer pairs for 25 ML models that cover crucial aspects of the model,
such as its training configurations, datasets, biases, architecture details,
and training resources. We employ annotators to extract the answers from the
original paper. Further, we explore the capabilities of LMs in generating model
cards by answering questions. Our initial experiments with ChatGPT-3.5, LLaMa,
and Galactica showcase a significant gap in the understanding of research
papers by these aforementioned LMs as well as generating factual textual
responses. We posit that our dataset can be used to train models to automate
the generation of model cards from paper text and reduce human effort in the
model card curation process. The complete dataset is available on
https://osf.io/hqt7p/?view_only=3b9114e3904c4443bcd9f5c270158d37
Related papers
- Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration [90.41908331897639]
Large language models (LLMs) have significantly benefited from training on diverse, high-quality task-specific data.
We present a novel approach, ReverseGen, designed to automatically generate effective training samples.
arXiv Detail & Related papers (2024-10-22T06:43:28Z) - "What is the value of {templates}?" Rethinking Document Information Extraction Datasets for LLMs [19.07429412219697]
We present K2Q, a collection of five datasets converted from KIE to a prompt-response format using a plethora of bespoke templates.
We empirically compare the performance of seven baseline generative models on K2Q with zero-shot prompting.
We find that creating diverse and intricate KIE questions enhances the performance and robustness of VRDU models.
arXiv Detail & Related papers (2024-10-20T19:42:30Z) - EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models [36.576853882830896]
We introduce EvolveDirector to train a text-to-image generation model comparable to advanced models using publicly available resources.
This framework interacts with advanced models through their public APIs to obtain text-image data pairs to train a base model.
We leverage pre-trained large vision-language models (VLMs) to guide the evolution of the base model.
arXiv Detail & Related papers (2024-10-09T17:52:28Z) - Molmo and PixMo: Open Weights and Open Data for State-of-the-Art Multimodal Models [146.18107944503436]
Molmo is a new family of VLMs that are state-of-the-art in their class of openness.
Our key innovation is a novel, highly detailed image caption dataset collected entirely from human annotators.
We will be releasing all of our model weights, captioning and fine-tuning data, and source code in the near future.
arXiv Detail & Related papers (2024-09-25T17:59:51Z) - Accelerating Large Language Model Pretraining via LFR Pedagogy: Learn, Focus, and Review [50.78587571704713]
Large Language Model (LLM) pretraining traditionally relies on autoregressive language modeling on randomly sampled data blocks from web-scale datasets.
We take inspiration from human learning techniques like spaced repetition to hypothesize that random data sampling for LLMs leads to high training cost and low quality models which tend to forget data.
In order to effectively commit web-scale information to long-term memory, we propose the LFR (Learn, Focus, and Review) pedagogy.
arXiv Detail & Related papers (2024-09-10T00:59:18Z) - ModelGPT: Unleashing LLM's Capabilities for Tailored Model Generation [35.160964210941955]
We propose ModelGPT, a framework designed to determine and generate AI models tailored to the data or task descriptions provided by the user.
Given user requirements, ModelGPT is able to provide tailored models at most 270x faster than the previous paradigms.
arXiv Detail & Related papers (2024-02-18T11:24:34Z) - What's documented in AI? Systematic Analysis of 32K AI Model Cards [40.170354637778345]
We conduct a comprehensive analysis of 32,111 AI model documentations on Hugging Face.
Most of the AI models with substantial downloads provide model cards, though the cards have uneven informativeness.
We find that sections addressing environmental impact, limitations, and evaluation exhibit the lowest filled-out rates, while the training section is the most consistently filled-out.
arXiv Detail & Related papers (2024-02-07T18:04:32Z) - Scalable Extraction of Training Data from (Production) Language Models [93.7746567808049]
This paper studies extractable memorization: training data that an adversary can efficiently extract by querying a machine learning model without prior knowledge of the training dataset.
We show an adversary can extract gigabytes of training data from open-source language models like Pythia or GPT-Neo, semi-open models like LLaMA or Falcon, and closed models like ChatGPT.
arXiv Detail & Related papers (2023-11-28T18:47:03Z) - GSAP-NER: A Novel Task, Corpus, and Baseline for Scholarly Entity
Extraction Focused on Machine Learning Models and Datasets [3.9169112083667073]
In academic writing, references to machine learning models and datasets are fundamental components.
Existing ground truth datasets do not treat fine-grained types like ML model and model architecture as separate entity types.
We release a corpus of 100 manually annotated full-text scientific publications and a first baseline model for 10 entity types centered around ML models and datasets.
arXiv Detail & Related papers (2023-11-16T12:43:02Z) - Delving Deeper into Data Scaling in Masked Image Modeling [145.36501330782357]
We conduct an empirical study on the scaling capability of masked image modeling (MIM) methods for visual recognition.
Specifically, we utilize the web-collected Coyo-700M dataset.
Our goal is to investigate how the performance changes on downstream tasks when scaling with different sizes of data and models.
arXiv Detail & Related papers (2023-05-24T15:33:46Z) - Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey [66.18478838828231]
Multi-modal pre-trained big models have drawn more and more attention in recent years.
This paper introduces the background of multi-modal pre-training by reviewing the conventional deep, pre-training works in natural language process, computer vision, and speech.
Then, we introduce the task definition, key challenges, and advantages of multi-modal pre-training models (MM-PTMs), and discuss the MM-PTMs with a focus on data, objectives, network, and knowledge enhanced pre-training.
arXiv Detail & Related papers (2023-02-20T15:34:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.