Bags of visual context-dependent words for generic visual categorization
First Claim
1. An image classification method comprising:
- generating a category context model for each of a plurality of image categories, the category context model for an image category including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in training images assigned to the category;
generating context information about an image to be classified, the context information including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in the image to be classified; and
assigning an image category to the image to be classified based on comparison of the context information about the image with the category context models for the image categories;
wherein at least the generating of the category context model and the generating of context information are performed by a computing apparatus.
6 Assignments
0 Petitions
Accused Products
Abstract
Category context models (64) and a universal context model (62) are generated including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in training images (50) assigned to each category and assigned to all categories, respectively. Context information (76) about an image to be classified (70) are generated including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in the image to be classified. For each category (82), a comparison is made of (i) closeness of the context information about the image to be classified with the corresponding category context model and (ii) closeness of the context information about the image to be classified with the universal context model. An image category (92) is assigned to the image to be classified being based on the comparisons.
-
Citations
15 Claims
-
1. An image classification method comprising:
-
generating a category context model for each of a plurality of image categories, the category context model for an image category including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in training images assigned to the category; generating context information about an image to be classified, the context information including sums of soft co-occurrences of pairs of visual words in geometric proximity to each other in the image to be classified; and assigning an image category to the image to be classified based on comparison of the context information about the image with the category context models for the image categories; wherein at least the generating of the category context model and the generating of context information are performed by a computing apparatus. - View Dependent Claims (2, 3)
-
-
4. An image classifier comprising:
-
a vocabulary of visual words wherein a visual word is a defined grouping of low-level image features; a patch context analyzer configured to generate a context representation for each of a plurality of patches of an image, wherein the context representation for a patch is based on occurrence probabilities of context visual words in a plurality of neighboring patches; an image labeler configured to assign an image category to an image based at least on the context representations of a plurality of patches of the image, wherein the image labeler applies a category context model for each image category that indicates probabilities of context words being in a neighborhood of an occurrence of a vocabulary word for images of that image category; and a category context model generator configured to generate each category context model as sums of soft co-occurrences of pairs of words in geometric proximity to each other in training images assigned to the category. - View Dependent Claims (5, 6, 7, 8, 10)
-
-
9. An image classifier comprising:
a vocabulary of visual words wherein a visual word is a defined grouping of low-level image features; a patch context analyzer configured to generate a context representation for each of a plurality of patches of an image, wherein the context representation for a patch is based on occurrence probabilities of context visual words in a plurality of neighboring patches; an image labeler configured to assign an image category to an image based at least on the context representations of a plurality of patches of the image, wherein the image labeler applies (i) a category context model for each image category that indicates probabilities of context words being in a neighborhood of an occurrence of a vocabulary word for images of that image category and (ii) a universal context model that indicates probabilities of context words being in a neighborhood of an occurrence of a vocabulary word for images regardless of image category; a category context model generator configured to generate each category context model as sums of soft co-occurrences of pairs of words in geometric proximity to each other in training images assigned to the category; and a universal context model generator configured to generate the universal context model as sums of soft co-occurrences of pairs of words in geometric proximity to each other in training images assigned to all categories. - View Dependent Claims (11, 12, 13, 14, 15)
Specification