Analyzing the Resource Utilization of Lambda Functions on Mobile Devices: Case Studies on Kotlin and Swift
- URL: http://arxiv.org/abs/2502.07809v1
- Date: Fri, 07 Feb 2025 20:26:07 GMT
- Title: Analyzing the Resource Utilization of Lambda Functions on Mobile Devices: Case Studies on Kotlin and Swift
- Authors: Chibundom U. Ejimuda, Gaston Longhitano, Reza Rawassizadeh,
- Abstract summary: Even minor reductions in smartphone power use could result in substantial energy savings.
This study explores the impact of Lambda functions on resource consumption in mobile programming.
- Score: 0.20482269513546458
- License:
- Abstract: With billions of smartphones in use globally, the daily time spent on these devices contributes significantly to overall electricity consumption. Given this scale, even minor reductions in smartphone power use could result in substantial energy savings. This study explores the impact of Lambda functions on resource consumption in mobile programming. While Lambda functions are known for enhancing code readability and conciseness, their use does not add to the functional capabilities of a programming language. Our research investigates the implications of using Lambda functions in terms of battery utilization, memory usage, and execution time compared to equivalent code structures without Lambda functions. Our findings reveal that Lambda functions impose a considerable resource overhead on mobile devices without offering additional functionalities.
Related papers
- Adaptive Tool Use in Large Language Models with Meta-Cognition Trigger [49.81945268343162]
We propose MeCo, an adaptive decision-making strategy for external tool use.
MeCo captures high-level cognitive signals in the representation space, guiding when to invoke tools.
Our experiments show that MeCo accurately detects LLMs' internal cognitive signals and significantly improves tool-use decision-making.
arXiv Detail & Related papers (2025-02-18T15:45:01Z) - ComplexFuncBench: Exploring Multi-Step and Constrained Function Calling under Long-Context Scenario [17.494787282066866]
We introduce ComplexFuncBench, a benchmark for complex function calling across five real-world scenarios.
Compared to existing benchmarks, ComplexFuncBench encompasses multi-step and constrained function calling.
We propose an automatic framework, ComplexEval, for quantitatively evaluating complex function calling tasks.
arXiv Detail & Related papers (2025-01-17T11:41:53Z) - Less is More: Optimizing Function Calling for LLM Execution on Edge Devices [0.44784055850794474]
Large Language Models (LLMs) struggle with function calling at the edge because they cannot handle complex inputs or manage multiple tools effectively.
We introduce Less-is-More, a novel fine-tuning-free function-calling scheme for dynamic tool selection.
Our approach is based on the key insight that selectively reducing the number of tools available to LLMs significantly improves their function-calling performance, execution time, and power efficiency on edge devices.
arXiv Detail & Related papers (2024-11-23T00:51:09Z) - Hammer: Robust Function-Calling for On-Device Language Models via Function Masking [26.495781685810044]
Hammer is a novel family of foundation models specifically engineered for on-device function calling.
Our empirical evaluations reveal that Hammer not only outperforms larger models but also demonstrates robust generalization across diverse benchmarks.
arXiv Detail & Related papers (2024-10-06T18:57:46Z) - ReLU$^2$ Wins: Discovering Efficient Activation Functions for Sparse
LLMs [91.31204876440765]
We introduce a general method that defines neuron activation through neuron output magnitudes and a tailored magnitude threshold.
To find the most efficient activation function for sparse computation, we propose a systematic framework.
We conduct thorough experiments on LLMs utilizing different activation functions, including ReLU, SwiGLU, ReGLU, and ReLU$2$.
arXiv Detail & Related papers (2024-02-06T08:45:51Z) - An LLM Compiler for Parallel Function Calling [68.04566807806071]
We introduce LLMCompiler, which executes functions in parallel to efficiently orchestrate multiple function calls.
We observe consistent latency speedup of up to 3.7x, cost savings of up to 6.7x, and accuracy improvement of up to 9% compared to ReAct.
arXiv Detail & Related papers (2023-12-07T18:32:04Z) - ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world
APIs [104.37772295581088]
Open-source large language models (LLMs), e.g., LLaMA, remain significantly limited in tool-use capabilities.
We introduce ToolLLM, a general tool-usetuning encompassing data construction, model training, and evaluation.
We first present ToolBench, an instruction-tuning framework for tool use, which is constructed automatically using ChatGPT.
arXiv Detail & Related papers (2023-07-31T15:56:53Z) - LongCoder: A Long-Range Pre-trained Language Model for Code Completion [56.813974784131624]
LongCoder employs a sliding window mechanism for self-attention and introduces two types of globally accessible tokens.
Bridge tokens are inserted throughout the input sequence to aggregate local information and facilitate global interaction.
memory tokens are included to highlight important statements that may be invoked later and need to be memorized.
arXiv Detail & Related papers (2023-06-26T17:59:24Z) - LambdaBeam: Neural Program Search with Higher-Order Functions and
Lambdas [44.919087759546635]
We design a search algorithm called LambdaBeam that can construct arbitrary functions that compose operations within a given DSL.
We train a neural policy network to choose which synthesiss to construct during search, and pass them as arguments to higher-order functions.
arXiv Detail & Related papers (2023-06-03T08:24:53Z) - LambdaNetworks: Modeling Long-Range Interactions Without Attention [3.459216990820884]
We present layers -- an alternative framework to self-attention -- for capturing long-range interactions between an input and structured contextual information.
Lambda layers capture such interactions by transforming available contexts into linear functions, termed COCOs.
They model both content and position-based interactions which enables their application to large structured inputs such as images.
arXiv Detail & Related papers (2021-02-17T06:33:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.