Title: Image Categorization by Learning and Reasoning with Regions
1Image Categorization by Learning and Reasoning
with Regions
- Yixin Chen, University of New Orleans
- James Z. Wang, The Pennsylvania State University
- Published on Journal of Machine Learning Research
5 (2004)
Presented by Jianhui Chen 09/26/2006
2Contents
- Introduction
- Image segmentation and representation
- An extension of Multiple Instance Learning
- - Diverse Density SVM (DD-SVM)
- Comparison between DD-SVM MI-SVM
- Experimental results
3Introduction
- Images (a) and (b) are mountains and glaciers
images. - Images (c), (d) and (e) are skiing images.
- Images (f) and (g) are beach images.
4Introduction
- A new learning technique for region-based image
categorization. - An extension from Multiple-Instance Learning
(MIL) DD-SVM.
5Image Segmentation and Representation
- Basic steps
- Step 1 -
- Divide the images into subblocks and extract
LUV features. - Step 2 -
- Do clustering using k-means and form regions.
- Step 3 -
- Form features vectors for regions (Classes).
-
6Image Segmentation and Representation
- Step 1
- (1) Partition the image into non-overlapping
- blocks of size 4 x 4 pixels.
- (2) Extract LUV features from each of the
- blocks, denoted as L, U, V.
-
7Image Segmentation and Representation
- Step 1
- (3) Apply Daubechies-4 wavelet transform and
compute features from - LH, HL, HH bands as fhl, flh and fhh .
- Suppose the coefficients are ckl,
ck,l1, ck1,l, cK1,l1, the feature is - computed as
(4) Form feature vector for each of the
subblocks as
L U V fhl flh fhh
8Image Segmentation and Representation
- Step 2
- (1) Apply k-means and do clustering.
- (2) Each resulting class corresponds to one
region.
9Image Segmentation and Representation
- Step 3
- (1) Compute the mean of the feature vectors
for each region. - (2) Compute the normalized inertia of order
1,2,3 for each region.
Normalized inertia
Shape feature of region Rj
Feature vector on Region Rj
10An Extension of Multiple Instance Learning
- Basic idea of DD-SVM
- (1) An images is referred as a bag which
- consists of instances.
- (2) Each bag is mapped to a point in a new
- feature space.
- (3) Standard SVMs are trained in the bag
- feature space.
11An Extension of Multiple Instance Learning
- Diverse-Density SVM (DD-SVM)
- (1) Maximum Margin Formulation of MIL.
- (2) Construct bag feature space based Diverse
- Density.
- (3) Compute region features vectors from
- instance features vectors.
- (4) A label is attached to a bag, instead of
- instances.
12An Extension of Multiple Instance Learning
- Objective function for DD-SVM
, define bag feature space.
, a kernel function.
C controls the trade-off between accuracy and
regularization.
13An Extension of Multiple Instance Learning
- The bag classifier is defined by as
Assume the bag feature space is given.
14An Extension of Multiple Instance Learning
- Constructing a Bag Feature Space
- (1) Diverse Density
- (2) Learning Instance Prototypes
- (3) Computing Bag Features
15Constructing a Bag Feature Space
x is a point in the instance feature space W is
a weight vector Ni is the number in the i-th bag
16Constructing a Bag Feature Space
- Learning Instance Prototypes
- (1) A large value of DD at a point indicates
it may fit better with the instances from
positive bags than with those from negative bags. - (2) Choose local maximizers as instance
prototypes. - (2) An instance prototype represents a class
of instances that is more likely to appear in
positive bags than in negative bags.
17Constructing a Bag Feature Space
- Learning Instance Prototype
18Constructing a Bag Feature Space
19Computing Bag features
(1) Each bag feature is defined by one instance
prototype and one instance from the bag. (2) A
bag feature gives the smallest distance between
any instance and the corresponding instance
prototype. (3) A bag feature can be viewed as a
measure of the degree that an instance prototype
shows up in the bag.
20Comparison between DD-SVM MI-SVM
- Learning process of DD-SVM
(1) Input is a collection of bags with binary
labels. (2) Output is SVM classifier.
21Comparison between DD-SVM MI-SVM
- Learning process of MI-SVM
22Comparison between DD-SVM MI-SVM
- Learning process of MI-SVM
- Input A collection of labeled bags.
- Output a SVM classifier.
-
23Comparison between DD-SVM MI-SVM
- DD-SVM
- (1) Negative bag all instances are negative
- (2) Positive bag at least one instance is
positive - MI-SVM
- (1) One instance is selected to represent the
whole positive bag. - (2) Train SVM using all of the negative
instances and selected positive instances.
24Experimental Results
- Experimental Setup and Data set
- (1) The data set consists of 2000 images from
20 image categories. - (2) All images are in JPEG format with the
size of 384 x 256 or 256 x 384. - (3) Manually set parameters, ie. C
- (4) Comparison among DD-SVM, MI-SVM
Hist-SVM
25Experimental Results
- Categorization Result in terms of accuracy
26Experimental Results
- Classification result in terms of confusion matrix
27Experimental Results
All Beach images contain mountains or
mountain-like regions All Mountain and glaciers
images contain regions corresponding to river,
lake or oceans.
28Experimental Results
- Sensitivity to image segmentation
29Experimental Results
- Sensitivity to the number of Categories in the
Data Set
30Experimental Results
- Sensitivity to the size and diversity of training
images
31Thank you !