An Ample Approach to Data and Modeling
- URL: http://arxiv.org/abs/2110.01776v1
- Date: Tue, 5 Oct 2021 01:26:09 GMT
- Title: An Ample Approach to Data and Modeling
- Authors: Luciano da F. Costa
- Abstract summary: We describe a framework for modeling how models can be built that integrates concepts and methods from a wide range of fields.
The reference M* meta model framework is presented, which relies critically in associating whole datasets and respective models in terms of a strict equivalence relation.
Several considerations about how the developed framework can provide insights about data clustering, complexity, collaborative research, deep learning, and creativity are then presented.
- Score: 1.0152838128195467
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In the present work, we describe a framework for modeling how models can be
built that integrates concepts and methods from a wide range of fields. The
information schism between the real-world and that which can be gathered and
considered by any individual information processing agent is characterized and
discussed, which is followed by the presentation of a series of the adopted
requisites while developing the modeling approach. The issue of mapping from
datasets into models is subsequently addressed, as well as some of the
respectively implied difficulties and limitations. Based on these
considerations, an approach to meta modeling how models are built is then
progressively developed. First, the reference M^* meta model framework is
presented, which relies critically in associating whole datasets and respective
models in terms of a strict equivalence relation. Among the interesting
features of this model are its ability to bridge the gap between data and
modeling, as well as paving the way to an algebra of both data and models which
can be employed to combine models into hierarchical manner. After illustrating
the M* model in terms of patterns derived from regular lattices, the reported
modeling approach continues by discussing how sampling issues, error and
overlooked data can be addressed, leading to the $M^{<\epsilon>}$ variant. The
situation in which the data needs to be represented in terms of respective
probability densities is treated next, yielding the $M^{<\sigma>}$ meta model,
which is then illustrated respectively to a real-world dataset (iris flowers
data). Several considerations about how the developed framework can provide
insights about data clustering, complexity, collaborative research, deep
learning, and creativity are then presented, followed by overall conclusions.
Related papers
- Consistent estimation of generative model representations in the data
kernel perspective space [13.099029073152257]
Generative models, such as large language models and text-to-image diffusion models, produce relevant information when presented a query.
Different models may produce different information when presented the same query.
We present novel theoretical results for embedding-based representations of generative models in the context of a set of queries.
arXiv Detail & Related papers (2024-09-25T19:35:58Z) - Has Your Pretrained Model Improved? A Multi-head Posterior Based
Approach [25.927323251675386]
We leverage the meta-features associated with each entity as a source of worldly knowledge and employ entity representations from the models.
We propose using the consistency between these representations and the meta-features as a metric for evaluating pre-trained models.
Our method's effectiveness is demonstrated across various domains, including models with relational datasets, large language models and image models.
arXiv Detail & Related papers (2024-01-02T17:08:26Z) - Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning [79.53130089003986]
Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
arXiv Detail & Related papers (2023-10-05T09:09:44Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Discriminative Multimodal Learning via Conditional Priors in Generative
Models [21.166519800652047]
This research studies the realistic scenario in which all modalities and class labels are available for model training.
We show, in this scenario, that the variational lower bound limits mutual information between joint representations and missing modalities.
arXiv Detail & Related papers (2021-10-09T17:22:24Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Unsupervised clustering of series using dynamic programming and neural
processes [0.0]
We would like to segment and cluster a series such that the resulting blocks present in each cluster are coherent with respect to a predefined model structure.
It is useful to establish a general framework that enables the integration of plausible models and also accommodates data-driven approach into one approximated model to assist the clustering task.
In this work, we investigate the use of neural processes to build the approximated model while yielding the same assumptions required by the algorithm presented in arXiv:2101.09512.
arXiv Detail & Related papers (2021-01-26T18:17:10Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Relating by Contrasting: A Data-efficient Framework for Multimodal
Generative Models [86.9292779620645]
We develop a contrastive framework for generative model learning, allowing us to train the model not just by the commonality between modalities, but by the distinction between "related" and "unrelated" multimodal data.
Under our proposed framework, the generative model can accurately identify related samples from unrelated ones, making it possible to make use of the plentiful unlabeled, unpaired multimodal data.
arXiv Detail & Related papers (2020-07-02T15:08:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.