Prof NB Venkateswarlu - PowerPoint PPT Presentation

1 / 62
About This Presentation
Title:

Prof NB Venkateswarlu

Description:

Two Aspects of Feature Extraction. Extracting useful features from images or any other measurements. ... Network Pruning and Rule Extraction. Network pruning ... – PowerPoint PPT presentation

Number of Views:122
Avg rating:3.0/5.0
Slides: 63
Provided by: vk21
Category:

less

Transcript and Presenter's Notes

Title: Prof NB Venkateswarlu


1
  • Prof NB Venkateswarlu
  • Head, IT, GVPCOE
  • Visakhapatnam
  • venkat_ritch_at_yahoo.com
  • www.ritchcenter.com/nbv

2
First Let me say Hearty Welcome to you All
3
  • Also, let me
  • congrachulate
  • Chairman,
  • Secretary/Correspondent

4
  • Principal,
  • Prof. Ravindra Babu
  • Vice-Principal

5
  • and other Organizers for planning for such a
    nice workshop with excellent themes.

6
My Talk
  • Feature Extraction/ Selection

7
A Typical Image Processing System contains
Image Acquisition
Image Pre-Processing
Image En-hancement
Image Seg-mentation
Image Featu-re Extraction
Image Class-fication
Image Unde-rstanding
8
Two Aspects of Feature Extraction
  • Extracting useful features from images or any
    other measurements.

9
  • Identifying Transformed Variables which are
    functions of original variables and having some
    charcateristics.

10
Feature Selection
  • Selecting Important Variables is Feature
    Selection

11
  • Some Features Used in I.P Applications

12
  • Shape based
  • Contour based
  • Area based
  • Transform based
  • Projections
  • Signature
  • Problem specific

13
Perimeter, length etc. First Convex hull is
extracted
14
Skeletons
15
(No Transcript)
16
Averaged Radial density
17
Radial Basis functions
18
Rose Plots
19
Chain Codes
20
Crack code - 32330300
21
Signature
22
Bending Energy
23
Chord Distribution
24
Fourier Descriptors
25
Structure
26
Splines
27
Horizontal and vertical projections
28
Elongatedness
29
Convex Hull
30
Compactness
31
RGB, R ,G and B bands
32
(No Transcript)
33
Classification/Pattern Recognition
  • Statistical
  • Syntactical Linguistic
  • Discriminant function
  • Fuzzy
  • Neural
  • Hybrid

34
Dimensionality Reduction
  • Feature selection (i.e., attribute subset
    selection)
  • Select a minimum set of features such that the
    probability distribution of different classes
    given the values for those features is as close
    as possible to the original distribution given
    the values of all features
  • reduce of patterns in the patterns, easier to
    understand
  • Heuristic methods (due to exponential of
    choices)
  • step-wise forward selection
  • step-wise backward elimination
  • combining forward selection and backward
    elimination
  • decision-tree induction

35
Example of Decision Tree Induction
  • Initial attribute set
  • A1, A2, A3, A4, A5, A6

36
Heuristic Feature Selection Methods
  • There are 2d possible sub-features of d features
  • Several heuristic feature selection methods
  • Best single features under the feature
    independence assumption choose by significance
    tests.
  • Best step-wise feature selection
  • The best single-feature is picked first
  • Then next best feature condition to the first,
    ...
  • Step-wise feature elimination
  • Repeatedly eliminate the worst feature
  • Best combined feature selection and elimination
  • Optimal branch and bound
  • Use feature elimination and backtracking

37
Why do We need?
  • A classifier performance depends on
  • No of features
  • Feature distinguishability
  • No of groups
  • Groups characteristics in multidimensional space.
  • Needed response time
  • Memory requirements

38
Feature Extraction Methods
  • We will find transformed variables which are
    functions of original variables.
  • A good example Though we may conduct tests in
    more than test (K-D), finally grading is done
    based on total marks (1-D)

39
Principal Component Analysis
  • Given N data vectors from k-dimensions, find c lt
    k orthogonal vectors that can be best used to
    represent data
  • The original data set is reduced to one
    consisting of N data vectors on c principal
    components (reduced dimensions)
  • Each data vector is a linear combination of the c
    principal component vectors
  • Works for numeric data only
  • Used when the number of dimensions is large

40
Principal Component Analysis
41
Principal Component Analysis
  • Aimed at finding new co-ordinate system which has
    some characteristics.
  • M4.5 4.25
  • Cov Matrix 2.57 1.86
  • 1.86 6.21
  • Eigen Values 6.99, 1.79
  • Eigen Vectors 0.387 0.922
  • -0.922 0.387

42
(No Transcript)
43
However in some cases it is not possible to have
PCA working.
44
Canonical Analysis
45
  • Unlike PCA which takes global mean and
    covariance, this takes between the group and
    within the group covariance matrix and the
    calculates canonical axes.

46
Standard Deviation A Simple Indicator
  • Correlation Coefficient

47
Feature Selection Group Separability Indices
48
(No Transcript)
49
(No Transcript)
50
(No Transcript)
51
(No Transcript)
52
Feature Selection Through Clustering
53
(No Transcript)
54
Selecting From 4 variables
55
Multi-Layer Perceptron
56
Network Pruning and Rule Extraction
  • Network pruning
  • Fully connected network will be hard to
    articulate
  • N input nodes, h hidden nodes and m output nodes
    lead to h(mN) weights
  • Pruning Remove some of the links without
    affecting classification accuracy of the network
  • Extracting rules from a trained network
  • Discretize activation values replace individual
    activation value by the cluster average
    maintaining the network accuracy
  • Enumerate the output from the discretized
    activation values to find rules between
    activation value and output
  • Find the relationship between the input and
    activation value
  • Combine the above two to have rules relating the
    output to input

57
Neural Networks for Feature Extraction
58
Self-organizing feature maps (SOMs)
  • Clustering is also performed by having several
    units competing for the current object
  • The unit whose weight vector is closest to the
    current object wins
  • The winner and its neighbors learn by having
    their weights adjusted
  • SOMs are believed to resemble processing that can
    occur in the brain
  • Useful for visualizing high-dimensional data in
    2- or 3-D space

59
Other Model-Based Clustering Methods
  • Neural network approaches
  • Represent each cluster as an exemplar, acting as
    a prototype of the cluster
  • New objects are distributed to the cluster whose
    exemplar is the most similar according to some
    dostance measure
  • Competitive learning
  • Involves a hierarchical architecture of several
    units (neurons)
  • Neurons compete in a winner-takes-all fashion
    for the object currently being presented

60
Model-Based Clustering Methods
61
SVM
  • SVM constructs nonlinear decision functions by
    training classifier to perform a linear
    separation in some high dimensional space which
    is nonlinearly related to the input space. A
    Mercer kernel is used for mapping.

62
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com