Parallel Performance-Energy Predictive Modeling of Browsers: Case Study
of Servo
- URL: http://arxiv.org/abs/2002.03850v1
- Date: Thu, 6 Feb 2020 20:16:14 GMT
- Title: Parallel Performance-Energy Predictive Modeling of Browsers: Case Study
of Servo
- Authors: Rohit Zambre, Lars Bergstrom, Laleh Aghababaie Beni, Aparna
Chandramowliswharan
- Abstract summary: We model the relationship between web page primitives and a web browser's parallel performance using supervised learning.
We consider energy usage trade-offs for different levels of performance improvements using automated labeling algorithms.
Experiments on a quad-core Intel Ivy Bridge laptop show that we can improve performance and energy usage by up to 94.52% and 46.32% respectively.
- Score: 0.9699640804685628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mozilla Research is developing Servo, a parallel web browser engine, to
exploit the benefits of parallelism and concurrency in the web rendering
pipeline. Parallelization results in improved performance for pinterest.com but
not for google.com. This is because the workload of a browser is dependent on
the web page it is rendering. In many cases, the overhead of creating,
deleting, and coordinating parallel work outweighs any of its benefits. In this
paper, we model the relationship between web page primitives and a web
browser's parallel performance using supervised learning. We discover a feature
space that is representative of the parallelism available in a web page and
characterize it using seven key features. Additionally, we consider energy
usage trade-offs for different levels of performance improvements using
automated labeling algorithms. Such a model allows us to predict the degree of
parallelism available in a web page and decide whether or not to render a web
page in parallel. This modeling is critical for improving the browser's
performance and minimizing its energy usage. We evaluate our model by using
Servo's layout stage as a case study. Experiments on a quad-core Intel Ivy
Bridge (i7-3615QM) laptop show that we can improve performance and energy usage
by up to 94.52% and 46.32% respectively on the 535 web pages considered in this
study. Looking forward, we identify opportunities to apply this model to other
stages of a browser's architecture as well as other performance- and
energy-critical devices.
Related papers
- Tur[k]ingBench: A Challenge Benchmark for Web Agents [41.852615330725655]
TurkingBench is a benchmark consisting of tasks presented as web pages with textual instructions and multi-modal contexts.
This benchmark includes 32.2K instances spread across 158 tasks.
We assess the performance of cutting-edge private and open-source models, including language-only and vision-language models.
arXiv Detail & Related papers (2024-03-18T16:06:30Z) - Green AI: A Preliminary Empirical Study on Energy Consumption in DL
Models Across Different Runtime Infrastructures [56.200335252600354]
It is common practice to deploy pre-trained models on environments distinct from their native development settings.
This led to the introduction of interchange formats such as ONNX, which includes its infrastructure, and ONNX, which work as standard formats.
arXiv Detail & Related papers (2024-02-21T09:18:44Z) - Anatomizing Deep Learning Inference in Web Browsers [17.63663828498732]
We make the first comprehensive performance measurement of in-browser inference to date.
Our approach proposes new metrics to measure in-browser inference: responsiveness, smoothness, and inference accuracy.
In-browser inference exhibits a substantial latency gap, averaging 16.9 times slower on CPU and 4.9 times slower on GPU compared to native inference on PC devices.
arXiv Detail & Related papers (2024-02-08T08:02:57Z) - Asynchronous Distributed Genetic Algorithms with Javascript and JSON [0.0]
We present a distributed evolutionary computation system that uses the computational capabilities of the ubiquituous web browser.
Since computing becomes a social activity and is inherently impredictable, in this paper we will explore the performance of this kind of virtual computer.
arXiv Detail & Related papers (2024-01-30T18:23:28Z) - Performance Tuning for GPU-Embedded Systems: Machine-Learning-based and
Analytical Model-driven Tuning Methodologies [0.0]
The study introduces an analytical model-driven tuning methodology and a Machine Learning (ML)-based tuning methodology.
We evaluate the performance of the two tuning methodologies for different parallel prefix implementations of the BPLG library in an NVIDIA Jetson system.
arXiv Detail & Related papers (2023-10-24T22:09:03Z) - Retentive Network: A Successor to Transformer for Large Language Models [91.6652200825638]
We propose Retentive Network (RetNet) as a foundation architecture for large language models.
We theoretically derive the connection between recurrence and attention.
Experimental results on language modeling show that RetNet achieves favorable scaling results, parallel training, low-cost deployment, and efficient inference.
arXiv Detail & Related papers (2023-07-17T16:40:01Z) - Multimodal Web Navigation with Instruction-Finetuned Foundation Models [99.14209521903854]
We study data-driven offline training for web agents with vision-language foundation models.
We propose an instruction-following multimodal agent, WebGUM, that observes both webpage screenshots and HTML pages.
We empirically demonstrate this recipe improves the agent's ability of grounded multimodal perception, HTML comprehension, and multi-step reasoning.
arXiv Detail & Related papers (2023-05-19T17:44:34Z) - Does compressing activations help model parallel training? [64.59298055364336]
We present the first empirical study on the effectiveness of compression methods for model parallelism.
We implement and evaluate three common classes of compression algorithms.
We evaluate these methods across more than 160 settings and 8 popular datasets.
arXiv Detail & Related papers (2023-01-06T18:58:09Z) - Benchopt: Reproducible, efficient and collaborative optimization
benchmarks [67.29240500171532]
Benchopt is a framework to automate, reproduce and publish optimization benchmarks in machine learning.
Benchopt simplifies benchmarking for the community by providing an off-the-shelf tool for running, sharing and extending experiments.
arXiv Detail & Related papers (2022-06-27T16:19:24Z) - Concept Extraction Using Pointer-Generator Networks [86.75999352383535]
We propose a generic open-domain OOV-oriented extractive model that is based on distant supervision of a pointer-generator network.
The model has been trained on a large annotated corpus compiled specifically for this task from 250K Wikipedia pages.
arXiv Detail & Related papers (2020-08-25T22:28:14Z) - Optimizing Streaming Parallelism on Heterogeneous Many-Core
Architectures: A Machine Learning Based Approach [16.702537371391053]
This article presents an automatic approach to derive a good solution for hardware resource partition and task granularity for task-based parallel applications on heterogeneous many-core architectures.
Our approach employs a performance model to estimate the resulting performance of the target application under a given resource partition and task granularity configuration.
Compared to the single-stream version, our approach achieves a 1.6x and 1.1x speedup on the XeonPhi and the GPU platform, respectively.
arXiv Detail & Related papers (2020-03-05T21:18:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.