There must be encapsulated nonconceptual content in vision
- URL: http://arxiv.org/abs/2503.15538v1
- Date: Thu, 06 Mar 2025 14:44:55 GMT
- Title: There must be encapsulated nonconceptual content in vision
- Authors: Vincent C. Müller,
- Abstract summary: I propose an argument to support Jerry Fodor's thesis that input systems are modular and thus informationally encapsulated.<n>It seems to follow that there is informationally encapsulated nonconceptual content in visual perception.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper I want to propose an argument to support Jerry Fodor's thesis (Fodor 1983) that input systems are modular and thus informationally encapsulated. The argument starts with the suggestion that there is a "grounding problem" in perception, i. e. that there is a problem in explaining how perception that can yield a visual experience is possible, how sensation can become meaningful perception of something for the subject. Given that visual experience is actually possible, this invites a transcendental argument that explains the conditions of its possibility. I propose that one of these conditions is the existence of a visual module in Fodor's sense that allows the step from sensation to object-identifying perception, thus enabling visual experience. It seems to follow that there is informationally encapsulated nonconceptual content in visual perception.
Related papers
- Wanting to Be Understood Explains the Meta-Problem of Consciousness [7.40601112616244]
We argue that such external representations are a pre-condition for access consciousness.<n>Our drive to be understood, and our low level sensorimotor capacities for grasping' so rich, that the demand for an explanation of the feel of experience cannot be satisfactory''
arXiv Detail & Related papers (2025-06-10T12:31:09Z) - Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts? [62.984473889987605]
We present a zero-shot framework for fine-grained visual concept learning by leveraging large language model and Visual Question Answering (VQA) system.
We pose these questions along with the query image to a VQA system and aggregate the answers to determine the presence or absence of an object in the test images.
Our experiments demonstrate comparable performance with existing zero-shot visual classification methods and few-shot concept learning approaches.
arXiv Detail & Related papers (2024-10-17T15:16:10Z) - Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMs [52.497823009176074]
Large Vision-Language Models (LVLMs) often produce responses that misalign with factual information, a phenomenon known as hallucinations.<n>We introduce Visual Description Grounded Decoding (VDGD), a training-free method designed to enhance visual perception and improve reasoning capabilities in LVLMs.
arXiv Detail & Related papers (2024-05-24T16:21:59Z) - Qualia and the Formal Structure of Meaning [0.0]
Empirically, subjective meaning is ubiquitous in conscious experiences.
Within the context of the mind-matter relation, we provide a formalization of subjective meaning associated to one's mental representations.
arXiv Detail & Related papers (2024-05-02T10:05:36Z) - Pre-trained Vision-Language Models Learn Discoverable Visual Concepts [33.302556000017844]
We aim to answer this question as visual concepts learned "for free" would enable wide applications.<n>We assume that the visual concepts, if captured by pre-trained VLMs, can be extracted by their vision-language interface with text-based concept prompts.<n>Our proposed concept discovery and learning framework is thus designed to identify a diverse list of generic visual concepts.
arXiv Detail & Related papers (2024-04-19T06:41:32Z) - How does the primate brain combine generative and discriminative
computations in vision? [4.691670689443386]
Two contrasting conceptions of the inference process have each been influential in research on biological vision and machine vision.
We show that vision inverts a generative model through an interrogation of the evidence in a process often thought to involve top-down predictions of sensory data.
We explain and clarify the terminology, review the key empirical evidence, and propose an empirical research program that transcends and sets the stage for revealing the mysterious hybrid algorithm of primate vision.
arXiv Detail & Related papers (2024-01-11T16:07:58Z) - Intrinsic Physical Concepts Discovery with Object-Centric Predictive
Models [86.25460882547581]
We introduce the PHYsical Concepts Inference NEtwork (PHYCINE), a system that infers physical concepts in different abstract levels without supervision.
We show that object representations containing the discovered physical concepts variables could help achieve better performance in causal reasoning tasks.
arXiv Detail & Related papers (2023-03-03T11:52:21Z) - Sources of Richness and Ineffability for Phenomenally Conscious States [57.8137804587998]
We provide an information theoretic dynamical systems perspective on the richness and ineffability of consciousness.
In our framework, the richness of conscious experience corresponds to the amount of information in a conscious state.
While our model may not settle all questions relating to the explanatory gap, it makes progress toward a fully physicalist explanation.
arXiv Detail & Related papers (2023-02-13T14:41:04Z) - Affection: Learning Affective Explanations for Real-World Visual Data [50.28825017427716]
We introduce and share with the research community a large-scale dataset that contains emotional reactions and free-form textual explanations for 85,007 publicly available images.
We show that there is significant common ground to capture potentially plausible emotional responses with a large support in the subject population.
Our work paves the way for richer, more human-centric, and emotionally-aware image analysis systems.
arXiv Detail & Related papers (2022-10-04T22:44:17Z) - Neuro-Symbolic Visual Reasoning: Disentangling "Visual" from "Reasoning" [49.76230210108583]
We propose a framework to isolate and evaluate the reasoning aspect of visual question answering (VQA) separately from its perception.
We also propose a novel top-down calibration technique that allows the model to answer reasoning questions even with imperfect perception.
On the challenging GQA dataset, this framework is used to perform in-depth, disentangled comparisons between well-known VQA models.
arXiv Detail & Related papers (2020-06-20T08:48:29Z) - Information-Theoretic Free Energy as Emotion Potential: Emotional
Valence as a Function of Complexity and Novelty [0.0]
We modeled arousal potential as information contents to be processed in the brain after sensory stimuli are perceived.
We demonstrated empirical evidence with visual stimuli supporting the hypothesis that the summation of perceived novelty and complexity shapes the inverse U-shaped beauty function.
arXiv Detail & Related papers (2020-03-23T04:10:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.