Textual-Based vs. Thinging Machines Conceptual Modeling
- URL: http://arxiv.org/abs/2506.02646v1
- Date: Tue, 03 Jun 2025 09:00:26 GMT
- Title: Textual-Based vs. Thinging Machines Conceptual Modeling
- Authors: Sabah Al-Fedaghi,
- Abstract summary: Software engineers typically interpret the domain description in natural language and translate it into a conceptual model.<n>Three approaches are used in this domain modeling: textual languages, diagrammatic languages, and a mixed based of text and diagrams.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Software engineers typically interpret the domain description in natural language and translate it into a conceptual model. Three approaches are used in this domain modeling: textual languages, diagrammatic languages, and a mixed based of text and diagrams. According to some researchers, relying on a diagrammatic notation levies certain burdens for designing large models because visual languages are intended to depict everything diagrammatically during a development process but fail to do so for a lack of developer efficiency. It is claimed that textual formats enable easier manipulation in editors and tools and facilitate the integration of ontologies in software systems. In this paper, we explore the problem of the relationship between textual format and diagramming in conceptual modeling. The main focus is modeling based on the so-called thinging machine (TM). Several examples are developed in detail to contrast side-by-side targeted domains represented in textual description and TM modeling. A TM model is defined as a thimac (thing/machine) with a time feature that forms dynamic events over static thimacs utilizing five generic actions: create, process, release, transfer, and receive. This provides a conceptual foundation that can be simplified further by eliminating the actions of release, transfer, and receive. A multilevel reduction in the TM diagram s complexity can also be achieved by assuming diagrammatic notations represent the actions of creation and processing. We envision that special tools will help improve developer efficiency. The study s results of contrasting textual and mix-based descriptions vs. TM modeling justify our claim that TM modeling is a more appropriate methodology than other diagrammatic schemes (e.g., UML classes) examined in this paper.
Related papers
- PreGenie: An Agentic Framework for High-quality Visual Presentation Generation [25.673526096069548]
PreGenie is an agentic and modular framework powered by multimodal large language models (MLLMs) for generating high-quality visual presentations.<n>It operates in two stages: (1) Analysis and Initial Generation, which summarizes multimodal input and generates initial code, and (2) Review and Re-generation, which iteratively reviews intermediate code and rendered slides to produce final, high-quality presentations.
arXiv Detail & Related papers (2025-05-27T18:36:19Z) - Decoupled Visual Interpretation and Linguistic Reasoning for Math Problem Solving [57.22004912994658]
Current large vision-language models (LVLMs) typically employ a connector module to link visual features with text embeddings of large language models (LLMs)<n>This paper proposes a paradigm shift: instead of training end-to-end vision-language reasoning models, we advocate for developing a decoupled reasoning framework.
arXiv Detail & Related papers (2025-05-23T08:18:00Z) - Establishing tool support for a concept DSL [0.0]
This thesis describes Conceptual, a DSL for modeling the behavior of software systems using self-contained and highly reusable units of concepts.<n>The suggested strategy is then implemented with a simple compiler, allowing developers to access and utilize Alloy's existing analysis tools for program reasoning.
arXiv Detail & Related papers (2025-03-07T09:18:31Z) - Thinging Machines for Requirements Engineering: Superseding Flowchart-Based Modeling [0.0]
It is claimed that present elicitation of requirements models focus on collecting information using natural language.<n>It is proposed that a solution to this problem involves using complexity theory, transdisciplinarity, multidimensionality and knowledge management.
arXiv Detail & Related papers (2025-01-28T05:30:45Z) - EditAR: Unified Conditional Generation with Autoregressive Models [58.093860528672735]
We propose EditAR, a single unified autoregressive framework for a variety of conditional image generation tasks.<n>The model takes both images and instructions as inputs, and predicts the edited images tokens in a vanilla next-token paradigm.<n>We evaluate its effectiveness across diverse tasks on established benchmarks, showing competitive performance to various state-of-the-art task-specific methods.
arXiv Detail & Related papers (2025-01-08T18:59:35Z) - Large Concept Models: Language Modeling in a Sentence Representation Space [62.73366944266477]
We present an attempt at an architecture which operates on an explicit higher-level semantic representation, which we name a concept.<n> Concepts are language- and modality-agnostic and represent a higher level idea or action in a flow.<n>We show that our model exhibits impressive zero-shot generalization performance to many languages.
arXiv Detail & Related papers (2024-12-11T23:36:20Z) - From Image to UML: First Results of Image Based UML Diagram Generation Using LLMs [1.961305559606562]
In software engineering processes, systems are first specified using a modeling language.
Large Language Models (LLM) are used to generate the formal representation of (UML) models from a given drawing.
More specifically, we have evaluated the capabilities of different LLMs to convert images of class diagrams into the actual models represented in the images.
arXiv Detail & Related papers (2024-04-17T13:33:11Z) - Divide and Conquer: Language Models can Plan and Self-Correct for
Compositional Text-to-Image Generation [72.6168579583414]
CompAgent is a training-free approach for compositional text-to-image generation with a large language model (LLM) agent as its core.
Our approach achieves more than 10% improvement on T2I-CompBench, a comprehensive benchmark for open-world compositional T2I generation.
arXiv Detail & Related papers (2024-01-28T16:18:39Z) - InstructCV: Instruction-Tuned Text-to-Image Diffusion Models as Vision Generalists [66.85125112199898]
We develop a unified language interface for computer vision tasks that abstracts away task-specific design choices.
Our model, dubbed InstructCV, performs competitively compared to other generalist and task-specific vision models.
arXiv Detail & Related papers (2023-09-30T14:26:43Z) - On Advances in Text Generation from Images Beyond Captioning: A Case
Study in Self-Rationalization [89.94078728495423]
We show that recent advances in each modality, CLIP image representations and scaling of language models, do not consistently improve multimodal self-rationalization of tasks with multimodal inputs.
Our findings call for a backbone modelling approach that can be built on to advance text generation from images and text beyond image captioning.
arXiv Detail & Related papers (2022-05-24T00:52:40Z) - Conceptual Software Engineering Applied to Movie Scripts and Stories [0.0]
This study focuses on conceptual modeling as a software engineering tool.
Researchers in the humanities and social sciences might not use the same degree of formalization as engineers.
The paper presents a novel approach to developing diagrammatic static/dynamic models of movie scripts and stories.
arXiv Detail & Related papers (2020-12-17T15:24:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.