SHAPED-BASED TECHNIQUES FOR EXPLORING DESIGN SPACES
First Claim
1. A computer-implemented method for generating computational representations for three-dimensional (3D) geometry shapes, the method comprising:
- for each view included in a first plurality of views associated with a first 3D geometry, generating a view activation based on a first convolutional neural network (CNN) block;
aggregating the view activations to generate a first tiled activation;
generating a first shape embedding having a fixed size based on the first tiled activation and a second CNN block;
generating a first plurality of re-constructed views based on the first shape embedding;
performing one or more training operations on at least one of the first CNN block or the second CNN block based on the first plurality of views and the first plurality of re-constructed views to generate a trained encoder; and
generating a second shape embedding having the fixed size based on the trained encoder.
1 Assignment
0 Petitions
Accused Products
Abstract
In various embodiments, a training application generates a trained encoder that automatically generates shape embeddings having a first size and representing three-dimensional (3D) geometry shapes. First, the training application generates a different view activation for each of multiple views associated with a first 3D geometry based on a first convolutional neural network (CNN) block. The training application then aggregates the view activations to generate a tiled activation. Subsequently, the training application generates a first shape embedding having the first size based on the tiled activation and a second CNN block. The training application then generates multiple re-constructed views based on the first shape embedding. The training application performs training operation(s) on at least one of the first CNN block and the second CNN block based on the views and the re-constructed views to generate the trained encoder.
-
Citations
20 Claims
-
1. A computer-implemented method for generating computational representations for three-dimensional (3D) geometry shapes, the method comprising:
-
for each view included in a first plurality of views associated with a first 3D geometry, generating a view activation based on a first convolutional neural network (CNN) block; aggregating the view activations to generate a first tiled activation; generating a first shape embedding having a fixed size based on the first tiled activation and a second CNN block; generating a first plurality of re-constructed views based on the first shape embedding; performing one or more training operations on at least one of the first CNN block or the second CNN block based on the first plurality of views and the first plurality of re-constructed views to generate a trained encoder; and generating a second shape embedding having the fixed size based on the trained encoder. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10)
-
-
11. One or more non-transitory computer readable media including instructions that, when executed by one or more processors, cause the one or more processors to perform the steps of:
-
for each view included in a first plurality of views associated with a first 3D geometry, generating a view activation based on a first convolutional neural network (CNN) block; aggregating the view activations to generate a first tiled activation; generating a first shape embedding having a fixed size based on the first tiled activation and a second CNN block; generating a first plurality of re-constructed views based on the first shape embedding; performing one or more training operations on at least one of the first CNN block or the second CNN block based on the first plurality of views and the first plurality of re-constructed views to generate a trained encoder; and generating a second shape embedding having the fixed size based on the trained encoder. - View Dependent Claims (12, 13, 14, 15, 16, 17, 18, 19)
-
-
20. A system, comprising:
-
one or more memories storing instructions; and one or more processors that are coupled to the one or more memories and, when executing the instructions, are configured to; for each view included in a first plurality of views associated with a first 3D geometry, generate a view activation based on a first convolutional neural network (CNN) block; aggregate the view activations to generate a first tiled activation; generate a first shape embedding having a fixed size based on the first tiled activation and a second CNN block; generate a first plurality of re-constructed views based on the first shape embedding; perform one or more training operations on at least one of the first CNN block or the second CNN block based on the first plurality of views and the first plurality of re-constructed views to generate a trained encoder; and generate a second shape embedding having the fixed size based on the trained encoder.
-
Specification