"Close...but not as good as an educator." -- Using ChatGPT to provide
formative feedback in large-class collaborative learning
- URL: http://arxiv.org/abs/2311.01634v1
- Date: Thu, 2 Nov 2023 23:00:38 GMT
- Title: "Close...but not as good as an educator." -- Using ChatGPT to provide
formative feedback in large-class collaborative learning
- Authors: Cory Dal Ponte, Sathana Dushyanthen and Kayley Lyons
- Abstract summary: We employed ChatGPT to provide personalised formative feedback in a one-hour Zoom break-out room activity.
Half of the 44 survey respondents had never used ChatGPT before.
Only three groups used the feedback loop to improve their evaluation plans.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Delivering personalised, formative feedback to multiple problem-based
learning groups in a short time period can be almost impossible. We employed
ChatGPT to provide personalised formative feedback in a one-hour Zoom break-out
room activity that taught practicing health professionals how to formulate
evaluation plans for digital health initiatives. Learners completed an
evaluation survey that included Likert scales and open-ended questions that
were analysed. Half of the 44 survey respondents had never used ChatGPT before.
Overall, respondents found the feedback favourable, described a wide range of
group dynamics, and had adaptive responses to the feedback, yet only three
groups used the feedback loop to improve their evaluation plans. Future
educators can learn from our experience including engineering prompts,
providing instructions on how to use ChatGPT, and scaffolding optimal group
interactions with ChatGPT. Future researchers should explore the influence of
ChatGPT on group dynamics and derive design principles for the use of ChatGPT
in collaborative learning.
Related papers
- Using ChatGPT for Science Learning: A Study on Pre-service Teachers'
Lesson Planning [0.7416846035207727]
This study analyzed lesson plans developed by 29 pre-service elementary teachers from a Korean university.
14 types of teaching and learning methods/strategies were identified in the lesson plans.
The study identified both appropriate and inappropriate use cases of ChatGPT in lesson planning.
arXiv Detail & Related papers (2024-01-18T22:52:04Z) - ChatGPT for Conversational Recommendation: Refining Recommendations by
Reprompting with Feedback [1.3654846342364308]
Large Language Models (LLMs) like ChatGPT have gained popularity due to their ease of use and their ability to adapt dynamically to various tasks while responding to feedback.
We build a rigorous pipeline around ChatGPT to simulate how a user might realistically probe the model for recommendations.
We explore the effect of popularity bias in ChatGPT's recommendations, and compare its performance to baseline models.
arXiv Detail & Related papers (2024-01-07T23:17:42Z) - Unreflected Acceptance -- Investigating the Negative Consequences of
ChatGPT-Assisted Problem Solving in Physics Education [4.014729339820806]
The impact of large language models (LLMs) on sensitive areas of everyday life, such as education, remains unclear.
Our work focuses on higher physics education and examines problem solving strategies.
arXiv Detail & Related papers (2023-08-21T16:14:34Z) - Transformative Effects of ChatGPT on Modern Education: Emerging Era of
AI Chatbots [36.760677949631514]
ChatGPT was released to provide coherent and useful replies based on analysis of large volumes of data.
Our preliminary evaluation concludes that ChatGPT performed differently in each subject area including finance, coding and maths.
There are clear drawbacks in its use, such as the possibility of producing inaccurate or false data.
Academic regulations and evaluation practices need to be updated, should ChatGPT be used as a tool in education.
arXiv Detail & Related papers (2023-05-25T17:35:57Z) - Exploring the Efficacy of ChatGPT in Analyzing Student Teamwork Feedback
with an Existing Taxonomy [0.0]
ChatGPT can achieve over 90% accuracy in labeling student comments.
This study contributes to the growing body of research on the use of AI models in educational contexts.
arXiv Detail & Related papers (2023-05-09T19:55:50Z) - ChatLog: Carefully Evaluating the Evolution of ChatGPT Across Time [54.18651663847874]
ChatGPT has achieved great success and can be considered to have acquired an infrastructural status.
Existing benchmarks encounter two challenges: (1) Disregard for periodical evaluation and (2) Lack of fine-grained features.
We construct ChatLog, an ever-updating dataset with large-scale records of diverse long-form ChatGPT responses for 21 NLP benchmarks from March, 2023 to now.
arXiv Detail & Related papers (2023-04-27T11:33:48Z) - ChatGPT Beyond English: Towards a Comprehensive Evaluation of Large
Language Models in Multilingual Learning [70.57126720079971]
Large language models (LLMs) have emerged as the most important breakthroughs in natural language processing (NLP)
This paper evaluates ChatGPT on 7 different tasks, covering 37 diverse languages with high, medium, low, and extremely low resources.
Compared to the performance of previous models, our extensive experimental results demonstrate a worse performance of ChatGPT for different NLP tasks and languages.
arXiv Detail & Related papers (2023-04-12T05:08:52Z) - To ChatGPT, or not to ChatGPT: That is the question! [78.407861566006]
This study provides a comprehensive and contemporary assessment of the most recent techniques in ChatGPT detection.
We have curated a benchmark dataset consisting of prompts from ChatGPT and humans, including diverse questions from medical, open Q&A, and finance domains.
Our evaluation results demonstrate that none of the existing methods can effectively detect ChatGPT-generated content.
arXiv Detail & Related papers (2023-04-04T03:04:28Z) - Is ChatGPT a Good NLG Evaluator? A Preliminary Study [121.77986688862302]
We provide a preliminary meta-evaluation on ChatGPT to show its reliability as an NLG metric.
Experimental results show that compared with previous automatic metrics, ChatGPT achieves state-of-the-art or competitive correlation with human judgments.
We hope our preliminary study could prompt the emergence of a general-purposed reliable NLG metric.
arXiv Detail & Related papers (2023-03-07T16:57:20Z) - Can ChatGPT Understand Too? A Comparative Study on ChatGPT and
Fine-tuned BERT [103.57103957631067]
ChatGPT has attracted great attention, as it can generate fluent and high-quality responses to human inquiries.
We evaluate ChatGPT's understanding ability by evaluating it on the most popular GLUE benchmark, and comparing it with 4 representative fine-tuned BERT-style models.
We find that: 1) ChatGPT falls short in handling paraphrase and similarity tasks; 2) ChatGPT outperforms all BERT models on inference tasks by a large margin; 3) ChatGPT achieves comparable performance compared with BERT on sentiment analysis and question answering tasks.
arXiv Detail & Related papers (2023-02-19T12:29:33Z) - Is ChatGPT a General-Purpose Natural Language Processing Task Solver? [113.22611481694825]
Large language models (LLMs) have demonstrated the ability to perform a variety of natural language processing (NLP) tasks zero-shot.
Recently, the debut of ChatGPT has drawn a great deal of attention from the natural language processing (NLP) community.
It is not yet known whether ChatGPT can serve as a generalist model that can perform many NLP tasks zero-shot.
arXiv Detail & Related papers (2023-02-08T09:44:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.