An Artificial Intelligence Browser Architecture (AIBA) For Our Kind and
Others: A Voice Name System Speech implementation with two warrants, Wake
Neutrality and Value Preservation of Privately Identifiable Information
- URL: http://arxiv.org/abs/2203.16497v1
- Date: Tue, 29 Mar 2022 11:49:41 GMT
- Title: An Artificial Intelligence Browser Architecture (AIBA) For Our Kind and
Others: A Voice Name System Speech implementation with two warrants, Wake
Neutrality and Value Preservation of Privately Identifiable Information
- Authors: Brian Subirana
- Abstract summary: Conversational commerce is the first of may applications based on always-on artificial intelligence systems that decide on its own when to interact with the environment.
Current dominant systems are closed garden solutions without wake neutrality and that can't fully exploit the PII data they have because of IRB and Cohues-type constraints.
We present a voice browser-and-server architecture that aims to address these two limitations by offering wake neutrality and the possibility to handle PII aiming to maximize its value.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conversational commerce, first pioneered by Apple's Siri, is the first of may
applications based on always-on artificial intelligence systems that decide on
its own when to interact with the environment, potentially collecting 24x7
longitudinal training data that is often Privately Identifiable Information
(PII). A large body of scholarly papers, on the order of a million according to
a simple Google Scholar search, suggests that the treatment of many health
conditions, including COVID-19 and dementia, can be vastly improved by this
data if the dataset is large enough as it has happened in other domains (e.g.
GPT3). In contrast, current dominant systems are closed garden solutions
without wake neutrality and that can't fully exploit the PII data they have
because of IRB and Cohues-type constraints.
We present a voice browser-and-server architecture that aims to address these
two limitations by offering wake neutrality and the possibility to handle PII
aiming to maximize its value. We have implemented this browser for the
collection of speech samples and have successfully demonstrated it can capture
over 200.000 samples of COVID-19 coughs. The architecture we propose is
designed so it can grow beyond our kind into other domains such as collecting
sound samples from vehicles, video images from nature, ingestible robotics,
multi-modal signals (EEG, EKG,...), or even interacting with other kinds such
as dogs and cats.
Related papers
- SONAR: A Synthetic AI-Audio Detection Framework and Benchmark [59.09338266364506]
SONAR is a synthetic AI-Audio Detection Framework and Benchmark.
It aims to provide a comprehensive evaluation for distinguishing cutting-edge AI-synthesized auditory content.
It is the first framework to uniformly benchmark AI-audio detection across both traditional and foundation model-based deepfake detection systems.
arXiv Detail & Related papers (2024-10-06T01:03:42Z) - Embodied-RAG: General Non-parametric Embodied Memory for Retrieval and Generation [65.23793829741014]
Embodied-RAG is a framework that enhances the model of an embodied agent with a non-parametric memory system.
At its core, Embodied-RAG's memory is structured as a semantic forest, storing language descriptions at varying levels of detail.
We demonstrate that Embodied-RAG effectively bridges RAG to the robotics domain, successfully handling over 200 explanation and navigation queries.
arXiv Detail & Related papers (2024-09-26T21:44:11Z) - MindSpeech: Continuous Imagined Speech Decoding using High-Density fNIRS and Prompt Tuning for Advanced Human-AI Interaction [0.0]
This paper reports a novel method for human-AI interaction by developing a direct brain-AI interface.
We discuss a novel AI model, called MindSpeech, which enables open-vocabulary, continuous decoding for imagined speech.
We demonstrate significant improvements in key metrics, such as BLEU-1 and BERT P scores, for three out of four participants.
arXiv Detail & Related papers (2024-07-25T16:39:21Z) - Consent in Crisis: The Rapid Decline of the AI Data Commons [74.68176012363253]
General-purpose artificial intelligence (AI) systems are built on massive swathes of public web data.
We conduct the first, large-scale, longitudinal audit of the consent protocols for the web domains underlying AI training corpora.
arXiv Detail & Related papers (2024-07-20T16:50:18Z) - Agent AI: Surveying the Horizons of Multimodal Interaction [83.18367129924997]
"Agent AI" is a class of interactive systems that can perceive visual stimuli, language inputs, and other environmentally-grounded data.
We envision a future where people can easily create any virtual reality or simulated scene and interact with agents embodied within the virtual environment.
arXiv Detail & Related papers (2024-01-07T19:11:18Z) - ProcTHOR: Large-Scale Embodied AI Using Procedural Generation [55.485985317538194]
ProcTHOR is a framework for procedural generation of Embodied AI environments.
We demonstrate state-of-the-art results across 6 embodied AI benchmarks for navigation, rearrangement, and arm manipulation.
arXiv Detail & Related papers (2022-06-14T17:09:35Z) - A Machine Learning-based Approach to Detect Threats in Bio-Cyber DNA
Storage Systems [20.27498894606937]
We have proposed an automated archival architecture which uses bioengineered bacteria to store and retrieve data, previously encoded into DNA.
The similarities between these biological media and classical ones can also be a drawback, as malicious parties might replicate traditional attacks on the former archival system.
In this paper, first we analyse the main characteristics of our storage system and the different types of attacks that could be executed on it.
Then, aiming at identifying on-going attacks, we propose and evaluate detection techniques, which rely on traditional metrics and machine learning algorithms.
arXiv Detail & Related papers (2020-09-28T14:55:20Z) - Deep Speaker Embeddings for Far-Field Speaker Recognition on Short
Utterances [53.063441357826484]
Speaker recognition systems based on deep speaker embeddings have achieved significant performance in controlled conditions.
Speaker verification on short utterances in uncontrolled noisy environment conditions is one of the most challenging and highly demanded tasks.
This paper presents approaches aimed to achieve two goals: a) improve the quality of far-field speaker verification systems in the presence of environmental noise, reverberation and b) reduce the system qualitydegradation for short utterances.
arXiv Detail & Related papers (2020-02-14T13:34:33Z) - Unsupervised Learning of Audio Perception for Robotics Applications:
Learning to Project Data to T-SNE/UMAP space [2.8935588665357077]
This paper builds upon key ideas to build perception of touch sounds without access to any ground-truth data.
We show how we can leverage ideas from classical signal processing to get large amounts of data of any sound of interest with a high precision.
arXiv Detail & Related papers (2020-02-10T20:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.