Virtual Mouse And Assistant: A Technological Revolution Of Artificial
Intelligence
- URL: http://arxiv.org/abs/2303.06309v1
- Date: Sat, 11 Mar 2023 05:00:06 GMT
- Title: Virtual Mouse And Assistant: A Technological Revolution Of Artificial
Intelligence
- Authors: Jagbeer Singh, Yash Goel, Shubhi Jain, Shiva Yadav
- Abstract summary: The purpose of this paper is to enhance the performance of the virtual assistant.
Virtual assistants can complete practically any specific smartphone or PC activity that you can complete on your own.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The purpose of this paper is to enhance the performance of the virtual
assistant. So, what exactly is a virtual assistant. Application software, often
called virtual assistants, also known as AI assistants or digital assistants,
is software that understands natural language voice commands and can perform
tasks on your behalf. What does a virtual assistant do. Virtual assistants can
complete practically any specific smartphone or PC activity that you can
complete on your own, and the list is continually expanding. Virtual assistants
typically do an impressive variety of tasks, including scheduling meetings,
delivering messages, and monitoring the weather. Previous virtual assistants,
like Google Assistant and Cortana, had limits in that they could only perform
searches and were not entirely automated. For instance, these engines do not
have the ability to forward and rewind the song in order to maintain the
control function of the song; they can only have the module to search for songs
and play them. Currently, we are working on a project where we are automating
Google, YouTube, and many other new things to improve the functionality of this
project. Now, in order to simplify the process, we've added a virtual mouse
that can only be used for cursor control and clicking. It receives input from
the camera, and our index finger acts as the mouse tip, our middle finger as
the right click, and so forth.
Related papers
- PIVOT: Iterative Visual Prompting Elicits Actionable Knowledge for VLMs [140.14239499047977]
Vision language models (VLMs) have shown impressive capabilities across a variety of tasks, from logical reasoning to visual understanding.
We propose a novel visual prompting approach for VLMs that we call Prompting with Iterative Visual Optimization (PIVOT)
We find, perhaps surprisingly, that our approach enables zero-shot control of robotic systems without any robot training data, navigation in a variety of environments, and other capabilities.
arXiv Detail & Related papers (2024-02-12T18:33:47Z) - Intelligent Virtual Assistants with LLM-based Process Automation [31.275267197246595]
This paper proposes a novel LLM-based virtual assistant that can automatically perform multi-step operations within mobile apps based on high-level user requests.
The system represents an advance in assistants by providing an end-to-end solution for parsing instructions, reasoning about goals, and executing actions.
arXiv Detail & Related papers (2023-12-04T07:51:58Z) - Voice-Based Smart Assistant System for Vehicles using RASA [0.0]
This paper focuses on the development of a voice-based smart assistance application for vehicles based on the RASA framework.
The smart assistant provides functionalities like navigation, communication via calls, getting weather forecasts and the latest news updates, and music that are completely voice-based in nature.
arXiv Detail & Related papers (2023-12-04T05:48:18Z) - HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging
Face [85.25054021362232]
Large language models (LLMs) have exhibited exceptional abilities in language understanding, generation, interaction, and reasoning.
LLMs could act as a controller to manage existing AI models to solve complicated AI tasks.
We present HuggingGPT, an LLM-powered agent that connects various AI models in machine learning communities.
arXiv Detail & Related papers (2023-03-30T17:48:28Z) - GRILLBot: An Assistant for Real-World Tasks with Neural Semantic Parsing
and Graph-Based Representations [5.545791216381869]
GRILLBot is the winning system in the 2022 Alexa Prize TaskBot Challenge.
It is a voice assistant to guide users through complex real-world tasks in the domains of cooking and home improvement.
arXiv Detail & Related papers (2022-08-31T14:24:35Z) - VRKitchen2.0-IndoorKit: A Tutorial for Augmented Indoor Scene Building
in Omniverse [77.52012928882928]
INDOORKIT is a built-in toolkit for NVIDIA OMNIVERSE.
It provides flexible pipelines for indoor scene building, scene randomizing, and animation controls.
arXiv Detail & Related papers (2022-06-23T17:53:33Z) - Play it by Ear: Learning Skills amidst Occlusion through Audio-Visual
Imitation Learning [62.83590925557013]
We learn a set of challenging partially-observed manipulation tasks from visual and audio inputs.
Our proposed system learns these tasks by combining offline imitation learning from tele-operated demonstrations and online finetuning.
In a set of simulated tasks, we find that our system benefits from using audio, and that by using online interventions we are able to improve the success rate of offline imitation learning by 20%.
arXiv Detail & Related papers (2022-05-30T04:52:58Z) - The MIT Voice Name System [53.473846742702854]
We aim to standardize voice interactions to a universal reach similar to that of other systems such as phone numbering.
We focus on voice as a starting point to talk to any IoT object.
Privacy and security are key elements considered because of speech-to-text errors and the amount of personal information contained in a voice sample.
arXiv Detail & Related papers (2022-03-28T19:09:26Z) - "Alexa, what do you do for fun?" Characterizing playful requests with
virtual assistants [68.48043257125678]
We introduce a taxonomy of playful requests rooted in theories of humor and refined by analyzing real-world traffic from Alexa.
Our conjecture is that understanding such utterances will improve user experience with virtual assistants.
arXiv Detail & Related papers (2021-05-12T10:48:00Z) - Implementation of Google Assistant & Amazon Alexa on Raspberry Pi [0.0]
This paper investigates the implementation of voice-enabled Google Assistant and Amazon Alexa on Raspberry Pi.
A voice-enabled system essentially means a system that processes voice as an input, decodes, or understands the meaning of that input and generates an appropriate voice output.
arXiv Detail & Related papers (2020-06-15T08:46:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.