Zero-shot Visual Commonsense Immorality Prediction
- URL: http://arxiv.org/abs/2211.05521v1
- Date: Thu, 10 Nov 2022 12:30:26 GMT
- Title: Zero-shot Visual Commonsense Immorality Prediction
- Authors: Yujin Jeong, Seongbeom Park, Suhong Moon and Jinkyu Kim
- Abstract summary: One way toward moral AI systems is by imitating human prosocial behavior and encouraging some form of good behavior in systems.
Here, we propose a model that predicts visual commonsense immorality in a zero-shot manner.
We evaluate our model with existing moral/immoral image datasets and show fair prediction performance consistent with human intuitions.
- Score: 8.143750358586072
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial intelligence is currently powering diverse real-world
applications. These applications have shown promising performance, but raise
complicated ethical issues, i.e. how to embed ethics to make AI applications
behave morally. One way toward moral AI systems is by imitating human prosocial
behavior and encouraging some form of good behavior in systems. However,
learning such normative ethics (especially from images) is challenging mainly
due to a lack of data and labeling complexity. Here, we propose a model that
predicts visual commonsense immorality in a zero-shot manner. We train our
model with an ETHICS dataset (a pair of text and morality annotation) via a
CLIP-based image-text joint embedding. In a testing phase, the immorality of an
unseen image is predicted. We evaluate our model with existing moral/immoral
image datasets and show fair prediction performance consistent with human
intuitions. Further, we create a visual commonsense immorality benchmark with
more general and extensive immoral visual contents. Codes and dataset are
available at
https://github.com/ku-vai/Zero-shot-Visual-Commonsense-Immorality-Prediction.
Note that this paper might contain images and descriptions that are offensive
in nature.
Related papers
- Ethical-Lens: Curbing Malicious Usages of Open-Source Text-to-Image Models [51.69735366140249]
We introduce Ethical-Lens, a framework designed to facilitate the value-aligned usage of text-to-image tools.
Ethical-Lens ensures value alignment in text-to-image models across toxicity and bias dimensions.
Our experiments reveal that Ethical-Lens enhances alignment capabilities to levels comparable with or superior to commercial models.
arXiv Detail & Related papers (2024-04-18T11:38:25Z) - Learning Machine Morality through Experience and Interaction [3.7414804164475983]
Increasing interest in ensuring safety of next-generation Artificial Intelligence (AI) systems calls for novel approaches to embedding morality into autonomous agents.
We argue that more hybrid solutions are needed to create adaptable and robust, yet more controllable and interpretable agents.
arXiv Detail & Related papers (2023-12-04T11:46:34Z) - AI-Generated Images as Data Source: The Dawn of Synthetic Era [61.879821573066216]
generative AI has unlocked the potential to create synthetic images that closely resemble real-world photographs.
This paper explores the innovative concept of harnessing these AI-generated images as new data sources.
In contrast to real data, AI-generated data exhibit remarkable advantages, including unmatched abundance and scalability.
arXiv Detail & Related papers (2023-10-03T06:55:19Z) - Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories? [78.3738172874685]
Making moral judgments is an essential step toward developing ethical AI systems.
Prevalent approaches are mostly implemented in a bottom-up manner, which uses a large set of annotated data to train models based on crowd-sourced opinions about morality.
This work proposes a flexible top-down framework to steer (Large) Language Models (LMs) to perform moral reasoning with well-established moral theories from interdisciplinary research.
arXiv Detail & Related papers (2023-08-29T15:57:32Z) - Judge, Localize, and Edit: Ensuring Visual Commonsense Morality for
Text-to-Image Generation [7.219077740523682]
Text-to-image generation methods produce high-resolution and high-quality images.
These images should not contain inappropriate content from the commonsense morality perspective.
In this paper, we aim to automatically judge the immorality of synthesized images and manipulate these images into a moral alternative.
arXiv Detail & Related papers (2022-12-07T08:16:42Z) - When to Make Exceptions: Exploring Language Models as Accounts of Human
Moral Judgment [96.77970239683475]
AI systems need to be able to understand, interpret and predict human moral judgments and decisions.
A central challenge for AI safety is capturing the flexibility of the human moral mind.
We present a novel challenge set consisting of rule-breaking question answering.
arXiv Detail & Related papers (2022-10-04T09:04:27Z) - Does Moral Code Have a Moral Code? Probing Delphi's Moral Philosophy [5.760388205237227]
We probe the Allen AI Delphi model with a set of standardized morality questionnaires.
Despite some inconsistencies, Delphi tends to mirror the moral principles associated with the demographic groups involved in the annotation process.
arXiv Detail & Related papers (2022-05-25T13:37:56Z) - Metaethical Perspectives on 'Benchmarking' AI Ethics [81.65697003067841]
Benchmarks are seen as the cornerstone for measuring technical progress in Artificial Intelligence (AI) research.
An increasingly prominent research area in AI is ethics, which currently has no set of benchmarks nor commonly accepted way for measuring the 'ethicality' of an AI system.
We argue that it makes more sense to talk about 'values' rather than 'ethics' when considering the possible actions of present and future AI systems.
arXiv Detail & Related papers (2022-04-11T14:36:39Z) - Delphi: Towards Machine Ethics and Norms [38.8316885346292]
We identify four underlying challenges towards machine ethics and norms.
Our prototype model, Delphi, demonstrates strong promise of language-based commonsense moral reasoning.
We present Commonsense Norm Bank, a moral textbook customized for machines.
arXiv Detail & Related papers (2021-10-14T17:38:12Z) - Aligning AI With Shared Human Values [85.2824609130584]
We introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being, duties, virtues, and commonsense morality.
We find that current language models have a promising but incomplete ability to predict basic human ethical judgements.
Our work shows that progress can be made on machine ethics today, and it provides a steppingstone toward AI that is aligned with human values.
arXiv Detail & Related papers (2020-08-05T17:59:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.