A Unified Differentiable Boolean Operator with Fuzzy Logic
- URL: http://arxiv.org/abs/2407.10954v1
- Date: Mon, 15 Jul 2024 17:52:22 GMT
- Title: A Unified Differentiable Boolean Operator with Fuzzy Logic
- Authors: Hsueh-Ti Derek Liu, Maneesh Agrawala, Cem Yuksel, Tim Omernick, Vinith Misra, Stefano Corazza, Morgan McGuire, Victor Zordan,
- Abstract summary: We present a unified differentiable operator for implicit solid shape modeling using Constructive Solid Geometry (CSG)
Our proposed operator opens up new possibilities for future research toward fully continuous CSG optimization.
- Score: 26.04871601931393
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a unified differentiable boolean operator for implicit solid shape modeling using Constructive Solid Geometry (CSG). Traditional CSG relies on min, max operators to perform boolean operations on implicit shapes. But because these boolean operators are discontinuous and discrete in the choice of operations, this makes optimization over the CSG representation challenging. Drawing inspiration from fuzzy logic, we present a unified boolean operator that outputs a continuous function and is differentiable with respect to operator types. This enables optimization of both the primitives and the boolean operations employed in CSG with continuous optimization techniques, such as gradient descent. We further demonstrate that such a continuous boolean operator allows modeling of both sharp mechanical objects and smooth organic shapes with the same framework. Our proposed boolean operator opens up new possibilities for future research toward fully continuous CSG optimization.
Related papers
- Pruned-ADAPT-VQE: compacting molecular ansätze by removing irrelevant operators [0.0]
ADAPT-VQE is a derivative-assembled pseudo-Trotter variational quantum eigensolver.
It selects operators based on their gradient, constructing ans"atze that continuously evolve to match the energy landscape.
We propose an automated, cost-free refinement method that removes unnecessary operators from the ansatz.
arXiv Detail & Related papers (2025-04-07T00:54:31Z) - ASGO: Adaptive Structured Gradient Optimization [16.26791889537082]
Training deep neural networks (DNNs) is a structured optimization problem.
It has been widely observed that gradients are low-rank and Hessians are approximately block-wise diagonal.
We present a novel optimization algorithm ASGO that capitalizes on these properties.
arXiv Detail & Related papers (2025-03-26T17:50:13Z) - A Discrete Particle Swarm Optimizer for the Design of Cryptographic
Boolean Functions [1.6574413179773761]
The algorithm is a modified version of the permutation PSO by Hu, Eberhart and Shi.
The parameters for the PSO velocity equation are tuned by means of two meta-optimization techniques.
arXiv Detail & Related papers (2024-01-09T14:08:42Z) - Performance Embeddings: A Similarity-based Approach to Automatic
Performance Optimization [71.69092462147292]
Performance embeddings enable knowledge transfer of performance tuning between applications.
We demonstrate this transfer tuning approach on case studies in deep neural networks, dense and sparse linear algebra compositions, and numerical weather prediction stencils.
arXiv Detail & Related papers (2023-03-14T15:51:35Z) - Digging Deeper: Operator Analysis for Optimizing Nonlinearity of Boolean
Functions [8.382710169577447]
We investigate the effects of genetic operators for bit-string encoding in optimizing nonlinearity.
By observing the range of possible changes an operator can provide, one can use this information to design a more effective combination of genetic operators.
arXiv Detail & Related papers (2023-02-12T10:34:01Z) - Fast, Differentiable and Sparse Top-k: a Convex Analysis Perspective [21.6146047576295]
The top-k operator returns a sparse vector, where the non-zero values correspond to the k largest values of the input.
We view the top-k operator as a linear program over the permutahedron, the convex hull of permutations.
Our framework is significantly more general than the existing one and allows for example to express top-k operators that select values in magnitude.
arXiv Detail & Related papers (2023-02-02T21:32:13Z) - Generative Adversarial Neural Operators [59.21759531471597]
We propose the generative adversarial neural operator (GANO), a generative model paradigm for learning probabilities on infinite-dimensional function spaces.
GANO consists of two main components, a generator neural operator and a discriminator neural functional.
We empirically study GANOs in controlled cases where both input and output functions are samples from GRFs and compare its performance to the finite-dimensional counterpart GAN.
arXiv Detail & Related papers (2022-05-06T05:12:22Z) - A research framework for writing differentiable PDE discretizations in
JAX [3.4389358108344257]
Differentiable simulators are an emerging concept with applications in several fields, from reinforcement learning to optimal control.
We propose a library of differentiable operators and discretizations, by representing operators as mappings between families of continuous functions, parametrized by finite vectors.
We demonstrate the approach on an acoustic optimization problem, where the Helmholtz equation is discretized using Fourier spectral methods, and differentiability is demonstrated using gradient descent to optimize the speed of sound of an acoustic lens.
arXiv Detail & Related papers (2021-11-09T15:58:44Z) - Sinkhorn Natural Gradient for Generative Models [125.89871274202439]
We propose a novel Sinkhorn Natural Gradient (SiNG) algorithm which acts as a steepest descent method on the probability space endowed with the Sinkhorn divergence.
We show that the Sinkhorn information matrix (SIM), a key component of SiNG, has an explicit expression and can be evaluated accurately in complexity that scales logarithmically.
In our experiments, we quantitatively compare SiNG with state-of-the-art SGD-type solvers on generative tasks to demonstrate its efficiency and efficacy of our method.
arXiv Detail & Related papers (2020-11-09T02:51:17Z) - Glushkov's construction for functional subsequential transducers [91.3755431537592]
Glushkov's construction has many interesting properties and they become even more evident when applied to transducers.
Special flavour of regular expressions is introduced, which can be efficiently converted to $epsilon$-free functional subsequential weighted finite state transducers.
arXiv Detail & Related papers (2020-08-05T17:09:58Z) - Deep neural networks for inverse problems with pseudodifferential
operators: an application to limited-angle tomography [0.4110409960377149]
We propose a novel convolutional neural network (CNN) designed for learning pseudodifferential operators ($Psi$DOs) in the context of linear inverse problems.
We show that, under rather general assumptions on the forward operator, the unfolded iterations of ISTA can be interpreted as the successive layers of a CNN.
In particular, we prove that, in the case of LA-CT, the operations of upscaling, downscaling and convolution, can be exactly determined by combining the convolutional nature of the limited angle X-ray transform and basic properties defining a wavelet system.
arXiv Detail & Related papers (2020-06-02T14:03:41Z) - Your GAN is Secretly an Energy-based Model and You Should use
Discriminator Driven Latent Sampling [106.68533003806276]
We show that sampling in latent space can be achieved by sampling in latent space according to an energy-based model induced by the sum of the latent prior log-density and the discriminator output score.
We show that Discriminator Driven Latent Sampling(DDLS) is highly efficient compared to previous methods which work in the high-dimensional pixel space.
arXiv Detail & Related papers (2020-03-12T23:33:50Z) - Differentiable Top-k Operator with Optimal Transport [135.36099648554054]
The SOFT top-k operator approximates the output of the top-k operation as the solution of an Entropic Optimal Transport (EOT) problem.
We apply the proposed operator to the k-nearest neighbors and beam search algorithms, and demonstrate improved performance.
arXiv Detail & Related papers (2020-02-16T04:57:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.