The Self 2.0: How AI-Enhanced Self-Clones Transform Self-Perception and
Improve Presentation Skills
- URL: http://arxiv.org/abs/2310.15112v1
- Date: Mon, 23 Oct 2023 17:20:08 GMT
- Title: The Self 2.0: How AI-Enhanced Self-Clones Transform Self-Perception and
Improve Presentation Skills
- Authors: Qingxiao Zheng, Yun Huang
- Abstract summary: This study explores the impact of AI-generated digital self-clones on improving online presentation skills.
We compared self-recorded videos (control) with self-clone videos (AI group) for English presentation practice.
Our findings recommend the ethical employment of digital self-clones to enhance the emotional and cognitive facets of skill development.
- Score: 9.495191491787908
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study explores the impact of AI-generated digital self-clones on
improving online presentation skills. We carried out a mixed-design experiment
involving 44 international students, comparing self-recorded videos (control)
with self-clone videos (AI group) for English presentation practice. The AI
videos utilized voice cloning, face swapping, lip-sync, and body-language
simulation to refine participants' original presentations in terms of
repetition, filler words, and pronunciation. Machine-rated scores indicated
enhancements in speech performance for both groups. Though the groups didn't
significantly differ, the AI group exhibited a heightened depth of reflection,
self-compassion, and a meaningful transition from a corrective to an enhancive
approach to self-critique. Within the AI group, congruence between
self-perception and AI self-clones resulted in diminished speech anxiety and
increased enjoyment. Our findings recommend the ethical employment of digital
self-clones to enhance the emotional and cognitive facets of skill development.
Related papers
- Transforming Higher Education with AI-Powered Video Lectures [0.2538209532048866]
The integration of artificial intelligence (AI) into video lecture production has the potential to transform higher education.<n>This paper investigates a semi automated workflow that combines Google Gemini for script generation, Amazon Polly for voice synthesis, and Microsoft PowerPoint for video assembly.
arXiv Detail & Related papers (2025-10-30T23:33:10Z) - Large Language Model Use Impact Locus of Control [0.0]
This paper explores the psychological impact of co-writing with AI on people's locus of control.<n>We found that employment status plays a critical role in shaping users' reliance on AI and their locus of control.
arXiv Detail & Related papers (2025-05-16T16:16:32Z) - Enhancing nonnative speech perception and production through an AI-powered application [0.0]
The aim of this study is to examine the impact of training with an AI-powered mobile application on nonnative sound perception and production.
The intervention involved training with the Speakometer mobile application, which incorporated recording tasks featuring the English vowels, along with pronunciation feedback and practice.
The results revealed significant improvements in both discrimination accuracy and production of the target contrast following the intervention.
arXiv Detail & Related papers (2025-03-18T10:05:12Z) - Identifying Features that Shape Perceived Consciousness in Large Language Model-based AI: A Quantitative Study of Human Responses [4.369058206183195]
This study quantitively examines which features of AI-generated text lead humans to perceive subjective consciousness in large language model (LLM)-based AI systems.
Using regression and clustering analyses, we investigated how these features influence participants' perceptions of AI consciousness.
arXiv Detail & Related papers (2025-02-21T10:27:28Z) - Adult learners recall and recognition performance and affective feedback when learning from an AI-generated synthetic video [1.7742433461734404]
The current study recruited 500 participants to investigate adult learners recall and recognition performances as well as their affective feedback on the AI-generated synthetic video.
The results indicated no statistically significant difference amongst conditions on recall and recognition performance.
However, adult learners preferred to learn from the video formats rather than text materials.
arXiv Detail & Related papers (2024-11-28T21:40:28Z) - Listening for Expert Identified Linguistic Features: Assessment of Audio Deepfake Discernment among Undergraduate Students [0.0]
This paper evaluates the impact of training undergraduate students to improve their audio deepfake discernment ability by listening for expert-defined linguistic features.
Our research goes beyond informational training by introducing targeted linguistic cues to listeners as a deepfake discernment mechanism.
Findings show that the experimental group showed a statistically significant decrease in their unsurety when evaluating audio clips and an improvement in their ability to correctly identify clips they were initially unsure about.
arXiv Detail & Related papers (2024-11-21T20:52:02Z) - MindSpeech: Continuous Imagined Speech Decoding using High-Density fNIRS and Prompt Tuning for Advanced Human-AI Interaction [0.0]
This paper reports a novel method for human-AI interaction by developing a direct brain-AI interface.
We discuss a novel AI model, called MindSpeech, which enables open-vocabulary, continuous decoding for imagined speech.
We demonstrate significant improvements in key metrics, such as BLEU-1 and BERT P scores, for three out of four participants.
arXiv Detail & Related papers (2024-07-25T16:39:21Z) - Collective Attention in Human-AI Teams [4.312803416185713]
The presence of an AI assistant significantly impacts team collective attention by modulating various aspects of shared cognition.
This study contributes to human-AI teaming research by highlighting collective attention as a central mechanism through which AI systems in team settings influence team performance.
arXiv Detail & Related papers (2024-07-03T13:46:00Z) - EgoNCE++: Do Egocentric Video-Language Models Really Understand Hand-Object Interactions? [48.702973928321946]
We introduce a novel asymmetric contrastive objective for EgoHOI named EgoNCE++.
Our experiments demonstrate that EgoNCE++ significantly boosts open-vocabulary HOI recognition, multi-instance retrieval, and action recognition tasks.
arXiv Detail & Related papers (2024-05-28T00:27:29Z) - Contrastive Learning of Person-independent Representations for Facial
Action Unit Detection [70.60587475492065]
We formulate the self-supervised AU representation learning signals in two-fold.
We contrast learn the AU representation within a video clip and devise a cross-identity reconstruction mechanism to learn the person-independent representations.
Our method outperforms other contrastive learning methods and significantly closes the performance gap between the self-supervised and supervised AU detection approaches.
arXiv Detail & Related papers (2024-03-06T01:49:28Z) - Attention-based Interactive Disentangling Network for Instance-level
Emotional Voice Conversion [81.1492897350032]
Emotional Voice Conversion aims to manipulate a speech according to a given emotion while preserving non-emotion components.
We propose an Attention-based Interactive diseNtangling Network (AINN) that leverages instance-wise emotional knowledge for voice conversion.
arXiv Detail & Related papers (2023-12-29T08:06:45Z) - DinoSR: Self-Distillation and Online Clustering for Self-supervised
Speech Representation Learning [140.96990096377127]
We introduce self-distillation and online clustering for self-supervised speech representation learning (DinoSR)
DinoSR first extracts contextualized embeddings from the input audio with a teacher network, then runs an online clustering system on the embeddings to yield a machine-discovered phone inventory, and finally uses the discretized tokens to guide a student network.
We show that DinoSR surpasses previous state-of-the-art performance in several downstream tasks, and provide a detailed analysis of the model and the learned discrete units.
arXiv Detail & Related papers (2023-05-17T07:23:46Z) - An Exploration of Self-Supervised Pretrained Representations for
End-to-End Speech Recognition [98.70304981174748]
We focus on the general applications of pretrained speech representations, on advanced end-to-end automatic speech recognition (E2E-ASR) models.
We select several pretrained speech representations and present the experimental results on various open-source and publicly available corpora for E2E-ASR.
arXiv Detail & Related papers (2021-10-09T15:06:09Z) - Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training [118.10946662410639]
We propose a novel Self-PU learning framework, which seamlessly integrates PU learning and self-training.
Self-PU highlights three "self"-oriented building blocks: a self-paced training algorithm that adaptively discovers and augments confident examples as the training proceeds.
We study a real-world application of PU learning, i.e., classifying brain images of Alzheimer's Disease.
arXiv Detail & Related papers (2020-06-22T17:53:59Z) - Does Visual Self-Supervision Improve Learning of Speech Representations
for Emotion Recognition? [63.564385139097624]
This work investigates visual self-supervision via face reconstruction to guide the learning of audio representations.
We show that a multi-task combination of the proposed visual and audio self-supervision is beneficial for learning richer features.
We evaluate our learned audio representations for discrete emotion recognition, continuous affect recognition and automatic speech recognition.
arXiv Detail & Related papers (2020-05-04T11:33:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.