ChatGPT & Mechanical Engineering: Examining performance on the FE
Mechanical Engineering and Undergraduate Exams
- URL: http://arxiv.org/abs/2309.15866v1
- Date: Tue, 26 Sep 2023 20:12:26 GMT
- Title: ChatGPT & Mechanical Engineering: Examining performance on the FE
Mechanical Engineering and Undergraduate Exams
- Authors: Matthew Frenkel, Hebah Emara
- Abstract summary: This study examines the capabilities of ChatGPT within the discipline of mechanical engineering.
It aims to examine use cases and pitfalls of such a technology in the classroom and professional settings.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The launch of ChatGPT at the end of 2022 generated large interest into
possible applications of artificial intelligence in STEM education and among
STEM professions. As a result many questions surrounding the capabilities of
generative AI tools inside and outside of the classroom have been raised and
are starting to be explored. This study examines the capabilities of ChatGPT
within the discipline of mechanical engineering. It aims to examine use cases
and pitfalls of such a technology in the classroom and professional settings.
ChatGPT was presented with a set of questions from junior and senior level
mechanical engineering exams provided at a large private university, as well as
a set of practice questions for the Fundamentals of Engineering Exam (FE) in
Mechanical Engineering. The responses of two ChatGPT models, one free to use
and one paid subscription, were analyzed. The paper found that the subscription
model (GPT-4) greatly outperformed the free version (GPT-3.5), achieving 76%
correct vs 51% correct, but the limitation of text only input on both models
makes neither likely to pass the FE exam. The results confirm findings in the
literature with regards to types of errors and pitfalls made by ChatGPT. It was
found that due to its inconsistency and a tendency to confidently produce
incorrect answers the tool is best suited for users with expert knowledge.
Related papers
- The potential of large language models for improving probability
learning: A study on ChatGPT3.5 and first-year computer engineering students [0.565395466029518]
ChatGPT is a large-scale language model that can solve probability problems.
ChatGPT is used in solving probability problems typically presented in computer engineering exams.
The model's ability to deliver high-quality explanations and illustrate solutions in any programming language suggests that large language models have the potential to serve as learning assistants.
arXiv Detail & Related papers (2023-10-09T12:54:58Z) - ChatGPT: A Study on its Utility for Ubiquitous Software Engineering
Tasks [2.084078990567849]
ChatGPT (Chat Generative Pre-trained Transformer) launched by OpenAI on November 30, 2022.
In this study, we explore how ChatGPT can be used to help with common software engineering tasks.
arXiv Detail & Related papers (2023-05-26T11:29:06Z) - ChatGPT-Crawler: Find out if ChatGPT really knows what it's talking
about [15.19126287569545]
This research examines the responses generated by ChatGPT from different Conversational QA corpora.
The study employed BERT similarity scores to compare these responses with correct answers and obtain Natural Language Inference(NLI) labels.
The study identified instances where ChatGPT provided incorrect answers to questions, providing insights into areas where the model may be prone to error.
arXiv Detail & Related papers (2023-04-06T18:42:47Z) - One Small Step for Generative AI, One Giant Leap for AGI: A Complete
Survey on ChatGPT in AIGC Era [95.2284704286191]
GPT-4 (a.k.a. ChatGPT plus) is one small step for generative AI (GAI) but one giant leap for artificial general intelligence (AGI)
Since its official release in November 2022, ChatGPT has quickly attracted numerous users with extensive media coverage.
This work is the first to survey ChatGPT with a comprehensive review of its underlying technology, applications, and challenges.
arXiv Detail & Related papers (2023-04-04T06:22:09Z) - To ChatGPT, or not to ChatGPT: That is the question! [78.407861566006]
This study provides a comprehensive and contemporary assessment of the most recent techniques in ChatGPT detection.
We have curated a benchmark dataset consisting of prompts from ChatGPT and humans, including diverse questions from medical, open Q&A, and finance domains.
Our evaluation results demonstrate that none of the existing methods can effectively detect ChatGPT-generated content.
arXiv Detail & Related papers (2023-04-04T03:04:28Z) - Can AI Chatbots Pass the Fundamentals of Engineering (FE) and Principles
and Practice of Engineering (PE) Structural Exams? [1.0554048699217669]
ChatGPT-4 and Bard, respectively, scored 70.9% and 39.2% in the FE exam and 46.2% and 41% in the PE exam.
It is evident that the current version of ChatGPT-4 could potentially pass the FE exam.
arXiv Detail & Related papers (2023-03-31T15:37:17Z) - On the Educational Impact of ChatGPT: Is Artificial Intelligence Ready
to Obtain a University Degree? [0.0]
We evaluate the influence of ChatGPT on university education.
We discuss how computer science higher education should adapt to tools like ChatGPT.
arXiv Detail & Related papers (2023-03-20T14:27:37Z) - Can ChatGPT Understand Too? A Comparative Study on ChatGPT and
Fine-tuned BERT [103.57103957631067]
ChatGPT has attracted great attention, as it can generate fluent and high-quality responses to human inquiries.
We evaluate ChatGPT's understanding ability by evaluating it on the most popular GLUE benchmark, and comparing it with 4 representative fine-tuned BERT-style models.
We find that: 1) ChatGPT falls short in handling paraphrase and similarity tasks; 2) ChatGPT outperforms all BERT models on inference tasks by a large margin; 3) ChatGPT achieves comparable performance compared with BERT on sentiment analysis and question answering tasks.
arXiv Detail & Related papers (2023-02-19T12:29:33Z) - BiasTestGPT: Using ChatGPT for Social Bias Testing of Language Models [73.29106813131818]
bias testing is currently cumbersome since the test sentences are generated from a limited set of manual templates or need expensive crowd-sourcing.
We propose using ChatGPT for the controllable generation of test sentences, given any arbitrary user-specified combination of social groups and attributes.
We present an open-source comprehensive bias testing framework (BiasTestGPT), hosted on HuggingFace, that can be plugged into any open-source PLM for bias testing.
arXiv Detail & Related papers (2023-02-14T22:07:57Z) - Is ChatGPT a General-Purpose Natural Language Processing Task Solver? [113.22611481694825]
Large language models (LLMs) have demonstrated the ability to perform a variety of natural language processing (NLP) tasks zero-shot.
Recently, the debut of ChatGPT has drawn a great deal of attention from the natural language processing (NLP) community.
It is not yet known whether ChatGPT can serve as a generalist model that can perform many NLP tasks zero-shot.
arXiv Detail & Related papers (2023-02-08T09:44:51Z) - A Categorical Archive of ChatGPT Failures [47.64219291655723]
ChatGPT, developed by OpenAI, has been trained using massive amounts of data and simulates human conversation.
It has garnered significant attention due to its ability to effectively answer a broad range of human inquiries.
However, a comprehensive analysis of ChatGPT's failures is lacking, which is the focus of this study.
arXiv Detail & Related papers (2023-02-06T04:21:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.