Hybrid unary-binary design for multiplier-less printed Machine Learning classifiers
- URL: http://arxiv.org/abs/2509.15316v1
- Date: Thu, 18 Sep 2025 18:02:24 GMT
- Title: Hybrid unary-binary design for multiplier-less printed Machine Learning classifiers
- Authors: Giorgos Armeniakos, Theodoros Mantzakidis, Dimitrios Soudris,
- Abstract summary: Printed Electronics (PE) provide a flexible, cost-efficient alternative to silicon for implementing machine learning (ML) circuits.<n>This work explores alternative arithmetic and a hybrid unary-binary architecture that removes costly encoders and enables efficient, multiplier-less execution of classifiers.<n> Evaluation on six datasets shows average reductions of 46% in area and 39% in power, with minimal accuracy loss, surpassing other state-of-the-art designs.
- Score: 3.0435742174040548
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Printed Electronics (PE) provide a flexible, cost-efficient alternative to silicon for implementing machine learning (ML) circuits, but their large feature sizes limit classifier complexity. Leveraging PE's low fabrication and NRE costs, designers can tailor hardware to specific ML models, simplifying circuit design. This work explores alternative arithmetic and proposes a hybrid unary-binary architecture that removes costly encoders and enables efficient, multiplier-less execution of MLP classifiers. We also introduce architecture-aware training to further improve area and power efficiency. Evaluation on six datasets shows average reductions of 46% in area and 39% in power, with minimal accuracy loss, surpassing other state-of-the-art MLP designs.
Related papers
- MAHL: Multi-Agent LLM-Guided Hierarchical Chiplet Design with Adaptive Debugging [61.83256382177746]
Large Language Models (LLMs) are promising to extend their abilities to 2.5D integration.<n>LLMs face challenges such as flatten design, high validation cost and imprecise parameter optimization.<n>We propose MAHL, a hierarchical LLM-based chiplet design generation framework.
arXiv Detail & Related papers (2025-08-08T05:47:31Z) - Investigating Structural Pruning and Recovery Techniques for Compressing Multimodal Large Language Models: An Empirical Study [64.26593350748401]
Multimodal Large Language Models (MLLMs) demonstrate impressive capabilities.<n>Current parameter reduction techniques primarily involve training MLLMs from Small Language Models (SLMs)<n>We propose to directly compress existing MLLMs through structural pruning combined with efficient recovery training.
arXiv Detail & Related papers (2025-07-28T11:57:52Z) - MiniCPM4: Ultra-Efficient LLMs on End Devices [126.22958722174583]
MiniCPM4 is a highly efficient large language model (LLM) designed explicitly for end-side devices.<n>We achieve this efficiency through systematic innovation in four key dimensions: model architecture, training data, training algorithms, and inference systems.
arXiv Detail & Related papers (2025-06-09T16:16:50Z) - Compact Yet Highly Accurate Printed Classifiers Using Sequential Support Vector Machine Circuits [0.6670927729669428]
We introduce the first sequential Support Vector Machine (SVM) classifiers.<n>Our SVMs yield on average 6x lower area and 4.6% higher accuracy compared to the printed state of the art.
arXiv Detail & Related papers (2025-02-03T16:30:27Z) - LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression Toolkit [55.73370804397226]
Quantization, a key compression technique, can effectively mitigate these demands by compressing and accelerating large language models.
We present LLMC, a plug-and-play compression toolkit, to fairly and systematically explore the impact of quantization.
Powered by this versatile toolkit, our benchmark covers three key aspects: calibration data, algorithms (three strategies), and data formats.
arXiv Detail & Related papers (2024-05-09T11:49:05Z) - Mechanistic Design and Scaling of Hybrid Architectures [114.3129802943915]
We identify and test new hybrid architectures constructed from a variety of computational primitives.
We experimentally validate the resulting architectures via an extensive compute-optimal and a new state-optimal scaling law analysis.
We find MAD synthetics to correlate with compute-optimal perplexity, enabling accurate evaluation of new architectures.
arXiv Detail & Related papers (2024-03-26T16:33:12Z) - Embedding Hardware Approximations in Discrete Genetic-based Training for Printed MLPs [1.4694098707981968]
Printed Electronics (PE) enables stretchable, conformal,and non-toxic hardware.
PE are constrained by larger feature sizes, making it challenging to implement complex circuits such as machine learning (ML)aware circuits.
In this paper, we maximize the benefits of approximate computing by integrating hardware approximation into the training process.
arXiv Detail & Related papers (2024-02-05T11:52:23Z) - Model-to-Circuit Cross-Approximation For Printed Machine Learning
Classifiers [4.865819809855699]
Printed electronics (PE) promises on-demand fabrication, low non-recurring engineering costs, and sub-cent fabrication costs.
Large feature sizes in PE prohibit the realization of complex ML models in PE, even with bespoke architectures.
We present an automated, cross-layer approximation framework tailored to bespoke architectures that enable complex ML models in PE.
arXiv Detail & Related papers (2023-03-14T22:11:34Z) - Learning a Fourier Transform for Linear Relative Positional Encodings in Transformers [71.32827362323205]
We propose a new class of linear Transformers calledLearner-Transformers (Learners)
They incorporate a wide range of relative positional encoding mechanisms (RPEs)
These include regular RPE techniques applied for sequential data, as well as novel RPEs operating on geometric data embedded in higher-dimensional Euclidean spaces.
arXiv Detail & Related papers (2023-02-03T18:57:17Z) - Effective Pre-Training Objectives for Transformer-based Autoencoders [97.99741848756302]
We study trade-offs between efficiency, cost and accuracy of Transformer encoders.
We combine features of common objectives and create new effective pre-training approaches.
arXiv Detail & Related papers (2022-10-24T18:39:44Z) - Approximate Decision Trees For Machine Learning Classification on Tiny
Printed Circuits [0.7349727826230862]
Printed Electronics (PE) cannot compete with silicon-based systems in conventional evaluation metrics.
PE offers attractive properties such as on-demand ultra-low-cost fabrication, flexibility and non-toxicity.
Despite the attractive characteristics of PE, the large feature sizes in PE prohibit the realization of complex printed circuits.
arXiv Detail & Related papers (2022-03-15T15:47:59Z) - Cross-Layer Approximation For Printed Machine Learning Circuits [4.865819809855699]
We propose and implement a cross-layer approximation, tailored for bespoke machine learning (ML) architectures in printed electronics (PE)
Our results demonstrate that our cross approximation delivers optimal designs that, compared to the state-of-the-art exact designs, feature 47% and 44% average area and power reduction, respectively, and less than 1% accuracy loss.
arXiv Detail & Related papers (2022-03-11T13:41:15Z) - Efficient pre-training objectives for Transformers [84.64393460397471]
We study several efficient pre-training objectives for Transformers-based models.
We prove that eliminating the MASK token and considering the whole output during the loss are essential choices to improve performance.
arXiv Detail & Related papers (2021-04-20T00:09:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.