Vanishing Point Detection with Direct and Transposed Fast Hough
Transform inside the neural network
- URL: http://arxiv.org/abs/2002.01176v3
- Date: Tue, 7 Jul 2020 13:08:55 GMT
- Title: Vanishing Point Detection with Direct and Transposed Fast Hough
Transform inside the neural network
- Authors: A. Sheshkus (4 and 6), A. Chirvonaya (2 and 6), D. Matveev (5 and 6),
D. Nikolaev (1 and 6), V.L. Arlazarov (3 and 4) ((1) Institute for
Information Transmission Problems (Kharkevich Institute) RAS, Moscow, Russia,
(2) National University of Science and Technology "MISIS", (3) Moscow
Institute for Physics and Technology, Moscow, Russia, (4) Institute for
Systems Analysis, Federal Research Center "Computer Science and Control" of
Russian Academy of Sciences, Moscow, Russia, (5) Lomonosov Moscow State
University, Moscow, Russia, (6) Smart Engines Service LLC, Moscow, Russia)
- Abstract summary: In this paper, we suggest a new neural network architecture for vanishing point detection in images.
The key element is the use of the direct and transposed Fast Hough Transforms separated by convolutional layer blocks with standard activation functions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we suggest a new neural network architecture for vanishing
point detection in images. The key element is the use of the direct and
transposed Fast Hough Transforms separated by convolutional layer blocks with
standard activation functions. It allows us to get the answer in the
coordinates of the input image at the output of the network and thus to
calculate the coordinates of the vanishing point by simply selecting the
maximum. Besides, it was proved that calculation of the transposed Fast Hough
Transform can be performed using the direct one. The use of integral operators
enables the neural network to rely on global rectilinear features in the image,
and so it is ideal for detecting vanishing points. To demonstrate the
effectiveness of the proposed architecture, we use a set of images from a DVR
and show its superiority over existing methods. Note, in addition, that the
proposed neural network architecture essentially repeats the process of direct
and back projection used, for example, in computed tomography.
Related papers
- HoughToRadon Transform: New Neural Network Layer for Features
Improvement in Projection Space [83.88591755871734]
HoughToRadon Transform layer is a novel layer designed to improve the speed of neural networks incorporated with Hough Transform.
Our experiments on the open MIDV-500 dataset show that this new approach leads to time savings and achieves state-of-the-art 97.7% accuracy.
arXiv Detail & Related papers (2024-02-05T12:19:16Z) - Implicit Neural Representation of Tileable Material Textures [1.1203075575217447]
We explore sinusoidal neural networks to represent periodic tileable textures.
We prove that the compositions of sinusoidal layers generate only integer frequencies with period $P$.
Our proposed neural implicit representation is compact and enables efficient reconstruction of high-resolution textures.
arXiv Detail & Related papers (2024-02-03T16:44:25Z) - Image segmentation with traveling waves in an exactly solvable recurrent
neural network [71.74150501418039]
We show that a recurrent neural network can effectively divide an image into groups according to a scene's structural characteristics.
We present a precise description of the mechanism underlying object segmentation in this network.
We then demonstrate a simple algorithm for object segmentation that generalizes across inputs ranging from simple geometric objects in grayscale images to natural images.
arXiv Detail & Related papers (2023-11-28T16:46:44Z) - In-Domain GAN Inversion for Faithful Reconstruction and Editability [132.68255553099834]
We propose in-domain GAN inversion, which consists of a domain-guided domain-regularized and a encoder to regularize the inverted code in the native latent space of the pre-trained GAN model.
We make comprehensive analyses on the effects of the encoder structure, the starting inversion point, as well as the inversion parameter space, and observe the trade-off between the reconstruction quality and the editing property.
arXiv Detail & Related papers (2023-09-25T08:42:06Z) - SPDER: Semiperiodic Damping-Enabled Object Representation [7.4297019016687535]
We present a neural network architecture designed to naturally learn a positional embedding.
The proposed architecture, SPDER, is a simple that uses an activation function composed of a sinusoidal multiplied by a sublinear function.
Our results indicate that SPDERs speed up training by 10x and converge to losses 1,500-50,000x lower than that of the state-of-the-art for image representation.
arXiv Detail & Related papers (2023-06-27T06:49:40Z) - Hamming Similarity and Graph Laplacians for Class Partitioning and
Adversarial Image Detection [2.960821510561423]
We investigate the potential for ReLU activation patterns (encoded as bit vectors) to aid in understanding and interpreting the behavior of neural networks.
We utilize Representational Dissimilarity Matrices (RDMs) to investigate the coherence of data within the embedding spaces of a deep neural network.
We demonstrate that bit vectors aid in adversarial image detection, again achieving over 95% accuracy in separating adversarial and non-adversarial images.
arXiv Detail & Related papers (2023-05-02T22:16:15Z) - AbHE: All Attention-based Homography Estimation [0.0]
We propose a strong-baseline model based on the Swin Transformer, which combines convolution neural network for local features and transformer module for global features.
In the homography regression stage, we adopt an attention layer for the channels of correlation volume, which can drop out some weak correlation feature points.
The experiment shows that in 8 Degree-of-Freedoms(DOFs) homography estimation our method overperforms the state-of-the-art method.
arXiv Detail & Related papers (2022-12-06T15:00:00Z) - Increasing the Accuracy of a Neural Network Using Frequency Selective
Mesh-to-Grid Resampling [4.211128681972148]
We propose the use of keypoint frequency selective mesh-to-grid resampling (FSMR) for the processing of input data for neural networks.
We show that depending on the network architecture and classification task the application of FSMR during training aids learning process.
The classification accuracy can be increased by up to 4.31 percentage points for ResNet50 and the Oxflower17 dataset.
arXiv Detail & Related papers (2022-09-28T21:34:47Z) - Weakly-supervised fire segmentation by visualizing intermediate CNN
layers [82.75113406937194]
Fire localization in images and videos is an important step for an autonomous system to combat fire incidents.
We consider weakly supervised segmentation of fire in images, in which only image labels are used to train the network.
We show that in the case of fire segmentation, which is a binary segmentation problem, the mean value of features in a mid-layer of classification CNN can perform better than conventional Class Activation Mapping (CAM) method.
arXiv Detail & Related papers (2021-11-16T11:56:28Z) - Deep Neural Networks are Surprisingly Reversible: A Baseline for
Zero-Shot Inversion [90.65667807498086]
This paper presents a zero-shot direct model inversion framework that recovers the input to the trained model given only the internal representation.
We empirically show that modern classification models on ImageNet can, surprisingly, be inverted, allowing an approximate recovery of the original 224x224px images from a representation after more than 20 layers.
arXiv Detail & Related papers (2021-07-13T18:01:43Z) - Spatially-Adaptive Pixelwise Networks for Fast Image Translation [57.359250882770525]
We introduce a new generator architecture, aimed at fast and efficient high-resolution image-to-image translation.
We use pixel-wise networks; that is, each pixel is processed independently of others.
Our model is up to 18x faster than state-of-the-art baselines.
arXiv Detail & Related papers (2020-12-05T10:02:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.