Black Box to White Box: Discover Model Characteristics Based on
Strategic Probing
- URL: http://arxiv.org/abs/2009.03136v1
- Date: Mon, 7 Sep 2020 14:44:28 GMT
- Title: Black Box to White Box: Discover Model Characteristics Based on
Strategic Probing
- Authors: Josh Kalin, Matthew Ciolino, David Noever, Gerry Dozier
- Abstract summary: White Box Adversarial Attacks rely on knowing underlying knowledge about the model attributes.
This works focuses on discovering to distrinct pieces of model information: the underlying architecture and primary training dataset.
With image classification, the focus is on exploring commonly deployed architectures and datasets available in popular public libraries.
Using a single transformer architecture with multiple levels of parameters, text generation is explored by fine tuning off different datasets.
Each dataset explored in image and text are distinguishable from one another.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In Machine Learning, White Box Adversarial Attacks rely on knowing underlying
knowledge about the model attributes. This works focuses on discovering to
distrinct pieces of model information: the underlying architecture and primary
training dataset. With the process in this paper, a structured set of input
probes and the output of the model become the training data for a deep
classifier. Two subdomains in Machine Learning are explored: image based
classifiers and text transformers with GPT-2. With image classification, the
focus is on exploring commonly deployed architectures and datasets available in
popular public libraries. Using a single transformer architecture with multiple
levels of parameters, text generation is explored by fine tuning off different
datasets. Each dataset explored in image and text are distinguishable from one
another. Diversity in text transformer outputs implies further research is
needed to successfully classify architecture attribution in text domain.
Related papers
- SegVG: Transferring Object Bounding Box to Segmentation for Visual Grounding [56.079013202051094]
We present SegVG, a novel method transfers the box-level annotation as signals to provide an additional pixel-level supervision for Visual Grounding.
This approach allows us to iteratively exploit the annotation as signals for both box-level regression and pixel-level segmentation.
arXiv Detail & Related papers (2024-07-03T15:30:45Z) - Multimodal Metadata Assignment for Cultural Heritage Artifacts [1.5826261914050386]
We develop a multimodal classifier for the cultural heritage domain using a late fusion approach.
The three modalities are Image, Text, and Tabular data.
All individual classifiers accurately predict missing properties in the digitized silk artifacts, with the multimodal approach providing the best results.
arXiv Detail & Related papers (2024-06-01T12:41:03Z) - Contrastive Transformer Learning with Proximity Data Generation for
Text-Based Person Search [60.626459715780605]
Given a descriptive text query, text-based person search aims to retrieve the best-matched target person from an image gallery.
Such a cross-modal retrieval task is quite challenging due to significant modality gap, fine-grained differences and insufficiency of annotated data.
In this paper, we propose a simple yet effective dual Transformer model for text-based person search.
arXiv Detail & Related papers (2023-11-15T16:26:49Z) - Memory Classifiers: Two-stage Classification for Robustness in Machine
Learning [19.450529711560964]
We propose a new method for classification which can improve robustness to distribution shifts.
We combine expert knowledge about the high-level" structure of the data with standard classifiers.
We show improvements which push beyond standard data augmentation techniques.
arXiv Detail & Related papers (2022-06-10T18:44:45Z) - Towards End-to-End Unified Scene Text Detection and Layout Analysis [60.68100769639923]
We introduce the task of unified scene text detection and layout analysis.
The first hierarchical scene text dataset is introduced to enable this novel research task.
We also propose a novel method that is able to simultaneously detect scene text and form text clusters in a unified way.
arXiv Detail & Related papers (2022-03-28T23:35:45Z) - Integrating Visuospatial, Linguistic and Commonsense Structure into
Story Visualization [81.26077816854449]
We first explore the use of constituency parse trees for encoding structured input.
Second, we augment the structured input with commonsense information and study the impact of this external knowledge on the generation of visual story.
Third, we incorporate visual structure via bounding boxes and dense captioning to provide feedback about the characters/objects in generated images.
arXiv Detail & Related papers (2021-10-21T00:16:02Z) - Pre-training Language Model Incorporating Domain-specific Heterogeneous Knowledge into A Unified Representation [49.89831914386982]
We propose a unified pre-trained language model (PLM) for all forms of text, including unstructured text, semi-structured text, and well-structured text.
Our approach outperforms the pre-training of plain text using only 1/4 of the data.
arXiv Detail & Related papers (2021-09-02T16:05:24Z) - Simple multi-dataset detection [83.9604523643406]
We present a simple method for training a unified detector on multiple large-scale datasets.
We show how to automatically integrate dataset-specific outputs into a common semantic taxonomy.
Our approach does not require manual taxonomy reconciliation.
arXiv Detail & Related papers (2021-02-25T18:55:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.