Decision Abstention Reduces Errors A Decision Abstaining Nversion Genetic Programming - PowerPoint PPT Presentation

1 / 17
About This Presentation
Title:

Decision Abstention Reduces Errors A Decision Abstaining Nversion Genetic Programming

Description:

The Initiative for Bioinformatics and Evolutionary Studies at the University of ... Inexpensive Beowulf cluster. Decision Abstention ... – PowerPoint PPT presentation

Number of Views:23
Avg rating:3.0/5.0
Slides: 18
Provided by: kosukei
Category:

less

Transcript and Presenter's Notes

Title: Decision Abstention Reduces Errors A Decision Abstaining Nversion Genetic Programming


1
Decision Abstention Reduces Errors A Decision
Abstaining N-version Genetic Programming
  • Kosuke Imamura Robert B. Heckendorn Terence
    Soule James A. Foster
  • The Initiative for Bioinformatics and
    Evolutionary Studies at the University of Idaho
    NIH NCRR grant 1P20RR016454-01 and NIH NCRR
    grant NIH NCRR 1P20RR016448-01 and NSF grant NSF
    EPS 809935.

2
What is the problem with GP?
  • Genetic Programming is unstable algorithm.
  • A trained individual produces faulty outputs.
  • Performance of equally fit individuals over the
    training data may widely vary on unseen data
    sets.
  • Consequently,
  • Given multiple equally high fit individuals,
    selecting one for actual use is a gamble.

3
The paper in a nutshell
  • Reducing errors of GP, and
  • Reducing performance fluctuations of GP

Proposal an ensemble of GP (N-version Genetic
Programming) with decision abstention
  • Questions addressed
  • What is an optimal ensemble GP?
  • When should a decision be abstained?

4
So, what is N-version Programming?
  • Correct output is C
  • Incorrect output isI
  • programA I C I C I C 3 faults
  • programB C C C I I C 2 faults
  • programC C I C C C I 2 faults
  • Result C C C C I C 1 fail
  • individual average faults2.3 (fault masking)

5
Then, our task is
  • Our task is to make sure that fault-masking
    occurs among individuals
  • Phenotypic diversity is a necessary condition.
    (disregard genotypic diversity)
  • Phenotypic diversity must quantifiably be defined.

6
What is the definition of diversity? A
Probabilistic Approach
  • Individuals must be reasonably high fit. (avoids
    combination of low fit individuals)
  • Independent Faults must be observed
    (quantifiable, individual learning).
  • Example if the fault rates of individuals are
    the same, then expected fault is under an area of
    a binomial probability density function

7
How do we find a probabilistically optimal
ensemble?
  • 1. Mass produce high fit individuals (we did it
    by an isolated island model on a cluster).
  • 2. Combine individuals to form an ensemble.
  • 3. Check if the error rate of the above ensemble
    is the expected error rate of independent faults.
  • 4. If the error rate is close enough to the
    expected rate, then done.
  • 5. Else form another ensemble and goto 2.

8
Contributions of NVGP
  • Defines the diversity in a quantifiable manner at
    a phenotypic level.
  • Provides a theoretically-backed-up evolution
    stopping criteria (optimal ensemble).
  • The proposed diversity quantification metric is
    applicable to other training based algorithms
    such as Neural Networks

9
An Idea Behind Abstention
  • Why cant a machine say, I dont know?
  • With abstention, a machine outputs,
  • Yes, No, and Dont know on a binary
    decision problem.

10
NVGP Demonstration problem (A Classification
Problem)
  • Ecoli DNA promoter region classification (a
    segment of DNA is a promoter region or not)
  • Implementation
  • - Linear Genome machines
  • - Isolated island model
  • - Inexpensive Beowulf cluster

11
Decision Abstention
  • A decision abstention occurs, when there is no
    decisive vote among the ensemble modules to make
    decision.
  • Unanimous vote is the most decisive
  • Tie vote is the least decisive
  • Needs ((N1)/2 h) votes
  • h is an abstention threshold

12
Results summary in two slide
Performance of NVGP alone
13
NVGP with decision abstention
50 error reduction
0 error
14
Effect of decision abstention
  • CorrectC IncorrectI Abstention threshold1
  • programA C C I C I C 2 faults
  • programB C I C I C C 2 faults
  • programC C I C I C I 3 faults
  • programD I C C I C C 2 faults
  • programE C I C C C I 2 faults
  • Majority vote C I C I C C 2 fail
  • Abstention C ? C ? C ? 0 fail

15
Trade-off between error reduction and abstention
rate
Adjusted Errors Q Ea ?N, Ea the number of
errors with abstention, N is the number of
dont know outputs, ? is a penalty
weight. Trade-off Q E0 (E0 number of errors by
simple majority)
(?0.5) abstention threshold test1 test2 test3 tes
t4 test5 0 6.7 8.0 10.1 7.7 6.8 1 7.2 8.3 10
.0 7.9 7.1 2 8.5 9.2 9.8 8.4 7.8 3 10.5 10.5
10.1 9.5 9.3
16
Conclusion
  • Abstention avoids random guesses.
  • High accuracy can be obtained at high abstention
    rate (Too much abstention makes the system of
    little use).
  • Abstention potentially indicates that the
    training set was not appropriate for particular
    instances.
  • For safety critical applications, a smaller ?
    value would be appropriate for the trade-off
    analysis. That is, do not penalize heavily when
    an ensemble is trying to avoid a random guess.

17
Future Research
  • Embed individual confidence
  • Thus, abstention occurs at both individual and
    ensemble bases.
Write a Comment
User Comments (0)
About PowerShow.com