From Far and Near: Perceptual Evaluation of Crowd Representations Across Levels of Detail
- URL: http://arxiv.org/abs/2510.20558v1
- Date: Thu, 23 Oct 2025 13:39:18 GMT
- Title: From Far and Near: Perceptual Evaluation of Crowd Representations Across Levels of Detail
- Authors: Xiaohan Sun, Carol O'Sullivan,
- Abstract summary: We investigate how users perceive the visual quality of crowd character representations at different levels of detail (LoD) and viewing distances.<n>Our results provide insights to guide the design of perceptually optimized LoD strategies for crowd rendering.
- Score: 1.0742675209112622
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we investigate how users perceive the visual quality of crowd character representations at different levels of detail (LoD) and viewing distances. Each representation: geometric meshes, image-based impostors, Neural Radiance Fields (NeRFs), and 3D Gaussians, exhibits distinct trade-offs between visual fidelity and computational performance. Our qualitative and quantitative results provide insights to guide the design of perceptually optimized LoD strategies for crowd rendering.
Related papers
- CoVis: A Collaborative Framework for Fine-grained Graphic Visual Understanding [0.29127054707887967]
CoVis is a collaborative framework for fine-grained visual understanding.<n>By designing and implementing a cascaded dual-layer segmentation network, it extracts as much knowledge as possible from an image.<n>It generates visual analytics for images, assisting observers in comprehending imagery from a more holistic perspective.
arXiv Detail & Related papers (2024-11-27T21:38:04Z) - Human Vision Constrained Super-Resolution [7.6097181983152185]
We propose an explicit Human Visual Processing Framework (HVPF) that guides SR methods according to human sensitivity to specific image details and viewing conditions.<n>We demonstrate the application of our framework in combination with network branching to improve the computational efficiency of SR methods.
arXiv Detail & Related papers (2024-11-26T15:24:45Z) - When Does Perceptual Alignment Benefit Vision Representations? [76.32336818860965]
We investigate how aligning vision model representations to human perceptual judgments impacts their usability.
We find that aligning models to perceptual judgments yields representations that improve upon the original backbones across many downstream tasks.
Our results suggest that injecting an inductive bias about human perceptual knowledge into vision models can contribute to better representations.
arXiv Detail & Related papers (2024-10-14T17:59:58Z) - Anti-Aliased Neural Implicit Surfaces with Encoding Level of Detail [54.03399077258403]
We present LoD-NeuS, an efficient neural representation for high-frequency geometry detail recovery and anti-aliased novel view rendering.
Our representation aggregates space features from a multi-convolved featurization within a conical frustum along a ray.
arXiv Detail & Related papers (2023-09-19T05:44:00Z) - Dehazed Image Quality Evaluation: From Partial Discrepancy to Blind
Perception [35.257798506356814]
Image dehazing aims to restore spatial details from hazy images.
We propose a Reduced-Reference dehazed image quality evaluation approach based on Partial Discrepancy.
We extend it to a No-Reference quality assessment metric with Blind Perception.
arXiv Detail & Related papers (2022-11-22T23:49:14Z) - Exploring CLIP for Assessing the Look and Feel of Images [87.97623543523858]
We introduce Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner.
Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments.
arXiv Detail & Related papers (2022-07-25T17:58:16Z) - Peripheral Vision Transformer [52.55309200601883]
We take a biologically inspired approach and explore to model peripheral vision in deep neural networks for visual recognition.
We propose to incorporate peripheral position encoding to the multi-head self-attention layers to let the network learn to partition the visual field into diverse peripheral regions given training data.
We evaluate the proposed network, dubbed PerViT, on the large-scale ImageNet dataset and systematically investigate the inner workings of the model for machine perception.
arXiv Detail & Related papers (2022-06-14T12:47:47Z) - PANet: Perspective-Aware Network with Dynamic Receptive Fields and
Self-Distilling Supervision for Crowd Counting [63.84828478688975]
We propose a novel perspective-aware approach called PANet to address the perspective problem.
Based on the observation that the size of the objects varies greatly in one image due to the perspective effect, we propose the dynamic receptive fields (DRF) framework.
The framework is able to adjust the receptive field by the dilated convolution parameters according to the input image, which helps the model to extract more discriminative features for each local region.
arXiv Detail & Related papers (2021-10-31T04:43:05Z) - Shallow Feature Based Dense Attention Network for Crowd Counting [103.67446852449551]
We propose a Shallow feature based Dense Attention Network (SDANet) for crowd counting from still images.
Our method outperforms other existing methods by a large margin, as is evident from a remarkable 11.9% Mean Absolute Error (MAE) drop of our SDANet.
arXiv Detail & Related papers (2020-06-17T13:34:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.