The trade-offs between Monolithic vs. Distributed Architectures
- URL: http://arxiv.org/abs/2405.03619v1
- Date: Mon, 6 May 2024 16:34:44 GMT
- Title: The trade-offs between Monolithic vs. Distributed Architectures
- Authors: Matheus Felisberto,
- Abstract summary: This article conducts a critical review of archi- tectural styles.
It focuses on the strengths and weaknesses of both monolithic and distributed architectures.
It also explores the role of cloud computing in transitioning from monolithic to distributed-based applications.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Software architects frequently engage in trade-off analysis, often confronting sub-optimal solutions due to unforeseen or overlooked disadvantages. Such outcomes can detrimentally affect a company's business operations and resource allocation. This article conducts a critical review of archi- tectural styles, particularly focusing on the strengths and weaknesses of both monolithic and distributed architectures, and their relationship to architectural characteristics. It also explores the role of cloud computing in transitioning from monolithic to distributed-based applications. Utilizing a broad range of sources, including papers and books from both industry and academia, this research provides an overview from theoretical foundations to practical applications. A notable trend observed is a shift back from distributed to monolithic architectures, possibly due to factors such as cost, complexity, and performance.
Related papers
- Unsupervised Graph Neural Architecture Search with Disentangled
Self-supervision [51.88848982611515]
Unsupervised graph neural architecture search remains unexplored in the literature.
We propose a novel Disentangled Self-supervised Graph Neural Architecture Search model.
Our model is able to achieve state-of-the-art performance against several baseline methods in an unsupervised manner.
arXiv Detail & Related papers (2024-03-08T05:23:55Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Serving Deep Learning Model in Relational Databases [72.72372281808694]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-artDL-Centric architecture offloadsDL computations to dedicated DL frameworks.
The potential UDF-Centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the database system.
The potentialRelation-Centric architecture aims to represent a large-scale tensor computation through operators.
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Data Mesh: a Systematic Gray Literature Review [3.038477115588261]
Data mesh is an emerging domain-driven decentralized data architecture that aims to minimize or avoid operational bottlenecks.
We systematically collected, analyzed, and synthesized 114 industrial gray literature articles.
The review provides insights into practitioners' perspectives on the four key principles of data mesh.
arXiv Detail & Related papers (2023-04-03T15:16:46Z) - A Theoretical Study of Inductive Biases in Contrastive Learning [32.98250585760665]
We provide the first theoretical analysis of self-supervised learning that incorporates the effect of inductive biases originating from the model class.
We show that when the model has limited capacity, contrastive representations would recover certain special clustering structures that are compatible with the model architecture.
arXiv Detail & Related papers (2022-11-27T01:53:29Z) - Visual Analysis of Neural Architecture Spaces for Summarizing Design
Principles [22.66053583920441]
ArchExplorer is a visual analysis method for understanding a neural architecture space and summarizing design principles.
A circle-packing-based architecture visualization has been developed to convey both the global relationships between clusters and local neighborhoods of the architectures in each cluster.
Two case studies and a post-analysis are presented to demonstrate the effectiveness of ArchExplorer in summarizing design principles and selecting better-performing architectures.
arXiv Detail & Related papers (2022-08-20T12:15:59Z) - On Neural Architecture Inductive Biases for Relational Tasks [76.18938462270503]
We introduce a simple architecture based on similarity-distribution scores which we name Compositional Network generalization (CoRelNet)
We find that simple architectural choices can outperform existing models in out-of-distribution generalizations.
arXiv Detail & Related papers (2022-06-09T16:24:01Z) - Dynamic enterprise architecture capabilities and organizational
benefits: an empirical mediation study [0.0]
This study focuses on EA-based capabilities, using the dynamic capabilities view as a theoretical foundation.
It develops and tests a new research model that explains how dynamic enterprise architecture capabilities lead to organizational benefits.
arXiv Detail & Related papers (2021-05-18T10:07:31Z) - Automated Search for Resource-Efficient Branched Multi-Task Networks [81.48051635183916]
We propose a principled approach, rooted in differentiable neural architecture search, to automatically define branching structures in a multi-task neural network.
We show that our approach consistently finds high-performing branching structures within limited resource budgets.
arXiv Detail & Related papers (2020-08-24T09:49:19Z) - Weight-Sharing Neural Architecture Search: A Battle to Shrink the
Optimization Gap [90.93522795555724]
Neural architecture search (NAS) has attracted increasing attentions in both academia and industry.
Weight-sharing methods were proposed in which exponentially many architectures share weights in the same super-network.
This paper provides a literature review on NAS, in particular the weight-sharing methods.
arXiv Detail & Related papers (2020-08-04T11:57:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.