A Systematic Literature Review of Empiricism and Norms of Reporting in
Computing Education Research Literature
- URL: http://arxiv.org/abs/2107.01984v1
- Date: Fri, 2 Jul 2021 16:37:29 GMT
- Title: A Systematic Literature Review of Empiricism and Norms of Reporting in
Computing Education Research Literature
- Authors: Sarah Heckman and Jeffrey C. Carver and Mark Sherriff and Ahmed
Al-Zubidy
- Abstract summary: The goal of this study is to characterize the reporting of empiricism in Computing Education Research (CER) literature.
We conducted an SLR of 427 papers published during 2014 and 2015 in five CER venues.
Over 80% of papers had some form of empirical evaluation.
- Score: 4.339510167603376
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computing Education Research (CER) is critical for supporting the increasing
number of students who need to learn computing skills. To systematically
advance knowledge, publications must be clear enough to support replications,
meta-analyses, and theory-building. The goal of this study is to characterize
the reporting of empiricism in CER literature by identifying whether
publications include information to support replications, meta-analyses, and
theory building. The research questions are: RQ1) What percentage of papers in
CER venues have empirical evaluation? RQ2) What are the characteristics of the
empirical evaluation? RQ3) Do the papers with empirical evaluation follow
reporting norms (both for inclusion and for labeling of key information)? We
conducted an SLR of 427 papers published during 2014 and 2015 in five CER
venues: SIGCSE TS, ICER, ITiCSE, TOCE, and CSE. We developed and applied the
CER Empiricism Assessment Rubric. Over 80% of papers had some form of empirical
evaluation. Quantitative evaluation methods were the most frequent. Papers most
frequently reported results on interventions around pedagogical techniques,
curriculum, community, or tools. There was a split in papers that had some type
of comparison between an intervention and some other data set or baseline. Many
papers lacked properly reported research objectives, goals, research questions,
or hypotheses, description of participants, study design, data collection, and
threats to validity. CER authors are contributing empirical results to the
literature; however, not all norms for reporting are met. We encourage authors
to provide clear, labeled details about their work so readers can use the
methodologies and results for replications and meta-analyses. As our community
grows, our reporting of CER should mature to help establish computing education
theory to support the next generation of computing learners.
Related papers
- Evaluating Human Alignment and Model Faithfulness of LLM Rationale [66.75309523854476]
We study how well large language models (LLMs) explain their generations through rationales.
We show that prompting-based methods are less "faithful" than attribution-based explanations.
arXiv Detail & Related papers (2024-06-28T20:06:30Z) - Categorical Syllogisms Revisited: A Review of the Logical Reasoning Abilities of LLMs for Analyzing Categorical Syllogism [62.571419297164645]
This paper provides a systematic overview of prior works on the logical reasoning ability of large language models for analyzing categorical syllogisms.
We first investigate all the possible variations for the categorical syllogisms from a purely logical perspective.
We then examine the underlying configurations (i.e., mood and figure) tested by the existing datasets.
arXiv Detail & Related papers (2024-06-26T21:17:20Z) - A Literature Review of Literature Reviews in Pattern Analysis and Machine Intelligence [58.6354685593418]
This paper proposes several article-level, field-normalized, and large language model-empowered bibliometric indicators to evaluate reviews.
The newly emerging AI-generated literature reviews are also appraised.
This work offers insights into the current challenges of literature reviews and envisions future directions for their development.
arXiv Detail & Related papers (2024-02-20T11:28:50Z) - Chain-of-Factors Paper-Reviewer Matching [32.86512592730291]
We propose a unified model for paper-reviewer matching that jointly considers semantic, topic, and citation factors.
We demonstrate the effectiveness of our proposed Chain-of-Factors model in comparison with state-of-the-art paper-reviewer matching methods and scientific pre-trained language models.
arXiv Detail & Related papers (2023-10-23T01:29:18Z) - Bias and Fairness in Large Language Models: A Survey [73.87651986156006]
We present a comprehensive survey of bias evaluation and mitigation techniques for large language models (LLMs)
We first consolidate, formalize, and expand notions of social bias and fairness in natural language processing.
We then unify the literature by proposing three intuitive, two for bias evaluation, and one for mitigation.
arXiv Detail & Related papers (2023-09-02T00:32:55Z) - Scientific Opinion Summarization: Paper Meta-review Generation Dataset, Methods, and Evaluation [55.00687185394986]
We propose the task of scientific opinion summarization, where research paper reviews are synthesized into meta-reviews.
We introduce the ORSUM dataset covering 15,062 paper meta-reviews and 57,536 paper reviews from 47 conferences.
Our experiments show that (1) human-written summaries do not always satisfy all necessary criteria such as depth of discussion, and identifying consensus and controversy for the specific domain, and (2) the combination of task decomposition and iterative self-refinement shows strong potential for enhancing the opinions.
arXiv Detail & Related papers (2023-05-24T02:33:35Z) - Artificial intelligence technologies to support research assessment: A
review [10.203602318836444]
This literature review identifies indicators that associate with higher impact or higher quality research from article text.
It includes studies that used machine learning techniques to predict citation counts or quality scores for journal articles or conference papers.
arXiv Detail & Related papers (2022-12-11T06:58:39Z) - Investigating Fairness Disparities in Peer Review: A Language Model
Enhanced Approach [77.61131357420201]
We conduct a thorough and rigorous study on fairness disparities in peer review with the help of large language models (LMs)
We collect, assemble, and maintain a comprehensive relational database for the International Conference on Learning Representations (ICLR) conference from 2017 to date.
We postulate and study fairness disparities on multiple protective attributes of interest, including author gender, geography, author, and institutional prestige.
arXiv Detail & Related papers (2022-11-07T16:19:42Z) - Best Practices and Scoring System on Reviewing A.I. based Medical
Imaging Papers: Part 1 Classification [0.9428556282541211]
The Machine Learning Education Sub-Committee of SIIM has identified a knowledge gap and a serious need to establish guidelines for reviewing these studies.
This first entry in the series focuses on the task of image classification.
The goal of this series is to provide resources to help improve the review process for A.I.-based medical imaging papers.
arXiv Detail & Related papers (2022-02-03T21:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.