Data-Driven Approach for Log Instruction Quality Assessment
- URL: http://arxiv.org/abs/2204.02618v1
- Date: Wed, 6 Apr 2022 07:02:23 GMT
- Title: Data-Driven Approach for Log Instruction Quality Assessment
- Authors: Jasmin Bogatinovski, Sasho Nedelkoski, Alexander Acker, Jorge Cardoso,
Odej Kao
- Abstract summary: There are no widely adopted guidelines on how to write log instructions with good quality properties.
We identify two quality properties: 1) correct log level assignment assessing the correctness of the log level, and 2) sufficient linguistic structure assessing the minimal richness of the static text necessary for verbose event description.
Our approach correctly assesses log level assignments with an accuracy of 0.88, and the sufficient linguistic structure with an F1 score of 0.99, outperforming the baselines.
- Score: 59.04636530383049
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the current IT world, developers write code while system operators run the
code mostly as a black box. The connection between both worlds is typically
established with log messages: the developer provides hints to the (unknown)
operator, where the cause of an occurred issue is, and vice versa, the operator
can report bugs during operation. To fulfil this purpose, developers write log
instructions that are structured text commonly composed of a log level (e.g.,
"info", "error"), static text ("IP {} cannot be reached"), and dynamic
variables (e.g. IP {}). However, as opposed to well-adopted coding practices,
there are no widely adopted guidelines on how to write log instructions with
good quality properties. For example, a developer may assign a high log level
(e.g., "error") for a trivial event that can confuse the operator and increase
maintenance costs. Or the static text can be insufficient to hint at a specific
issue. In this paper, we address the problem of log quality assessment and
provide the first step towards its automation. We start with an in-depth
analysis of quality log instruction properties in nine software systems and
identify two quality properties: 1) correct log level assignment assessing the
correctness of the log level, and 2) sufficient linguistic structure assessing
the minimal richness of the static text necessary for verbose event
description. Based on these findings, we developed a data-driven approach that
adapts deep learning methods for each of the two properties. An extensive
evaluation on large-scale open-source systems shows that our approach correctly
assesses log level assignments with an accuracy of 0.88, and the sufficient
linguistic structure with an F1 score of 0.99, outperforming the baselines. Our
study shows the potential of the data-driven methods in assessing instructions
quality and aid developers in comprehending and writing better code.
Related papers
- Studying and Benchmarking Large Language Models For Log Level Suggestion [49.176736212364496]
Large Language Models (LLMs) have become a focal point of research across various domains.
This paper investigates the impact of characteristics and learning paradigms on the performance of 12 open-source LLMs in log level suggestion.
arXiv Detail & Related papers (2024-10-11T03:52:17Z) - CraftRTL: High-quality Synthetic Data Generation for Verilog Code Models with Correct-by-Construction Non-Textual Representations and Targeted Code Repair [4.554742043916029]
This paper first presents an analysis of fine-tuned LLMs on Verilog coding, with synthetic data from prior methods.
We identify two main issues: difficulties in handling non-textual representations and significant variability during training with models randomly making "minor" mistakes.
Our fine-tuned Starcoder2-15B outperforms prior state-of-the-art results by 3.8%, 10.9%, 6.6% for pass@1 on VerilogEval-Machine, VerilogEval-Human, and RTLLM.
arXiv Detail & Related papers (2024-09-19T12:15:55Z) - LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Feedback [71.95402654982095]
We propose Math-Minos, a natural language feedback-enhanced verifier.
Our experiments reveal that a small set of natural language feedback can significantly boost the performance of the verifier.
arXiv Detail & Related papers (2024-06-20T06:42:27Z) - A Large-Scale Evaluation for Log Parsing Techniques: How Far Are We? [42.56249610409624]
We provide a new collection of annotated log datasets, denoted Loghub-2.0, which can better reflect the characteristics of log data in real-world software systems.
We conduct a thorough re-evaluation of 15 state-of-the-art logs in a more rigorous and practical setting. Particularly, we introduce a new evaluation metric to mitigate the sensitivity of existing metrics to imbalanced data distributions.
arXiv Detail & Related papers (2023-08-21T16:24:15Z) - Are They All Good? Studying Practitioners' Expectations on the
Readability of Log Messages [18.823475517909884]
Despite the importance of log messages, there is still a lack of standards on what constitutes good readability in log messages.
We conduct a series of interviews with 17 industrial practitioners to investigate their expectations on the readability of log messages.
We find that both deep learning and machine learning models can effectively classify the readability of log messages with a balanced accuracy above 80.0% on average.
arXiv Detail & Related papers (2023-08-17T07:53:24Z) - MURMUR: Modular Multi-Step Reasoning for Semi-Structured Data-to-Text
Generation [102.20036684996248]
We propose MURMUR, a neuro-symbolic modular approach to text generation from semi-structured data with multi-step reasoning.
We conduct experiments on two data-to-text generation tasks like WebNLG and LogicNLG.
arXiv Detail & Related papers (2022-12-16T17:36:23Z) - Borrowing from Similar Code: A Deep Learning NLP-Based Approach for Log
Statement Automation [0.0]
We introduce an updated and improved log-aware code-clone detection method to predict the location of logging statements.
We incorporate natural language processing (NLP) and deep learning methods to automate the log statements' description prediction.
Our analysis shows that our hybrid NLP and code-clone detection approach (NLP CC'd) outperforms conventional clone detectors in finding log statement locations.
arXiv Detail & Related papers (2021-12-02T14:03:49Z) - LogLAB: Attention-Based Labeling of Log Data Anomalies via Weak
Supervision [63.08516384181491]
We present LogLAB, a novel modeling approach for automated labeling of log messages without requiring manual work by experts.
Our method relies on estimated failure time windows provided by monitoring systems to produce precise labeled datasets in retrospect.
Our evaluation shows that LogLAB consistently outperforms nine benchmark approaches across three different datasets and maintains an F1-score of more than 0.98 even at large failure time windows.
arXiv Detail & Related papers (2021-11-02T15:16:08Z) - Self-Supervised Log Parsing [59.04636530383049]
Large-scale software systems generate massive volumes of semi-structured log records.
Existing approaches rely on log-specifics or manual rule extraction.
We propose NuLog that utilizes a self-supervised learning model and formulates the parsing task as masked language modeling.
arXiv Detail & Related papers (2020-03-17T19:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.