Non Linear Software Documentation with Interactive Code Examples
- URL: http://arxiv.org/abs/2311.18057v1
- Date: Wed, 29 Nov 2023 20:08:46 GMT
- Title: Non Linear Software Documentation with Interactive Code Examples
- Authors: Mathieu Nassif and Martin P. Robillard
- Abstract summary: Casdoc documents are interactive resources centered around code examples for programmers.
Explanations of the code elements are presented as annotations that the readers reveal based on their needs.
We observed that interactive documents can contain more information than static documents without being distracting to readers.
- Score: 9.880887106904519
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Documentation enables sharing knowledge between the developers of a
technology and its users. Creating quality documents, however, is challenging:
Documents must satisfy the needs of a large audience without being overwhelming
for individuals. We address this challenge with a new document format, named
Casdoc. Casdoc documents are interactive resources centered around code
examples for programmers. Explanations of the code elements are presented as
annotations that the readers reveal based on their needs. We evaluated Casdoc
in a field study with over 300 participants who used 126 documents as part of a
software design course. The majority of participants adopted Casdoc instead of
a baseline format during the study. We observed that interactive documents can
contain more information than static documents without being distracting to
readers. We also gathered insights into five aspects of Casdoc that can be
applied to other formats, and propose five guidelines to improve navigability
in online documents.
Related papers
- Unified Multi-Modal Interleaved Document Representation for Information Retrieval [57.65409208879344]
We produce more comprehensive and nuanced document representations by holistically embedding documents interleaved with different modalities.
Specifically, we achieve this by leveraging the capability of recent vision-language models that enable the processing and integration of text, images, and tables into a unified format and representation.
arXiv Detail & Related papers (2024-10-03T17:49:09Z) - DocGenome: An Open Large-scale Scientific Document Benchmark for Training and Testing Multi-modal Large Language Models [63.466265039007816]
We present DocGenome, a structured document benchmark constructed by annotating 500K scientific documents from 153 disciplines in the arXiv open-access community.
We conduct extensive experiments to demonstrate the advantages of DocGenome and objectively evaluate the performance of large models on our benchmark.
arXiv Detail & Related papers (2024-06-17T15:13:52Z) - Knowledge-Driven Cross-Document Relation Extraction [3.868708275322908]
Relation extraction (RE) is a well-known NLP application often treated as a sentence- or document-level task.
We propose a novel approach, KXDocRE, that embed domain knowledge of entities with input text for cross-document RE.
arXiv Detail & Related papers (2024-05-22T11:30:59Z) - Understanding Documentation Use Through Log Analysis: An Exploratory
Case Study of Four Cloud Services [14.104545948572836]
We analyze documentation page-view logs from four cloud-based industrial services.
By analyzing page-view logs for over 100,000 users, we find diverse patterns of documentation page visits.
We propose documentation page-view log analysis as a feasible technique for design audits of documentation.
arXiv Detail & Related papers (2023-10-16T20:37:29Z) - DocumentNet: Bridging the Data Gap in Document Pre-Training [78.01647768018485]
We propose a method to collect massive-scale and weakly labeled data from the web to benefit the training of VDER models.
The collected dataset, named DocumentNet, does not depend on specific document types or entity sets.
Experiments on a set of broadly adopted VDER tasks show significant improvements when DocumentNet is incorporated into the pre-training.
arXiv Detail & Related papers (2023-06-15T08:21:15Z) - Learning Diverse Document Representations with Deep Query Interactions
for Dense Retrieval [79.37614949970013]
We propose a new dense retrieval model which learns diverse document representations with deep query interactions.
Our model encodes each document with a set of generated pseudo-queries to get query-informed, multi-view document representations.
arXiv Detail & Related papers (2022-08-08T16:00:55Z) - Layout-Aware Information Extraction for Document-Grounded Dialogue:
Dataset, Method and Demonstration [75.47708732473586]
We propose a layout-aware document-level Information Extraction dataset, LIE, to facilitate the study of extracting both structural and semantic knowledge from visually rich documents.
LIE contains 62k annotations of three extraction tasks from 4,061 pages in product and official documents.
Empirical results show that layout is critical for VRD-based extraction, and system demonstration also verifies that the extracted knowledge can help locate the answers that users care about.
arXiv Detail & Related papers (2022-07-14T07:59:45Z) - Unified Pretraining Framework for Document Understanding [52.224359498792836]
We present UDoc, a new unified pretraining framework for document understanding.
UDoc is designed to support most document understanding tasks, extending the Transformer to take multimodal embeddings as input.
An important feature of UDoc is that it learns a generic representation by making use of three self-supervised losses.
arXiv Detail & Related papers (2022-04-22T21:47:04Z) - Synthetic Document Generator for Annotation-free Layout Recognition [15.657295650492948]
We describe a synthetic document generator that automatically produces realistic documents with labels for spatial positions, extents and categories of layout elements.
We empirically illustrate that a deep layout detection model trained purely on the synthetic documents can match the performance of a model that uses real documents.
arXiv Detail & Related papers (2021-11-11T01:58:44Z) - Modeling Endorsement for Multi-Document Abstractive Summarization [10.166639983949887]
A crucial difference between single- and multi-document summarization is how salient content manifests itself in the document(s)
In this paper, we model the cross-document endorsement effect and its utilization in multiple document summarization.
Our method generates a synopsis from each document, which serves as an endorser to identify salient content from other documents.
arXiv Detail & Related papers (2021-10-15T03:55:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.