H05ABF (PDF version)
H Chapter Contents
H Chapter Introduction
NAG Library Manual

NAG Library Routine Document

H05ABF

Note:  before using this routine, please read the Users' Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent details.

+ Contents

    1  Purpose
    7  Accuracy

1  Purpose

Given a set of m features and a scoring mechanism for any subset of those features, H05ABF selects the best n subsets of size p using a direct communication branch and bound algorithm.

2  Specification

SUBROUTINE H05ABF ( MINCR, M, IP, NBEST, LA, BSCORE, BZ, F, MINCNT, GAMMA, ACC, IUSER, RUSER, IFAIL)
INTEGER  MINCR, M, IP, NBEST, LA, BZ(M-IP,NBEST), MINCNT, IUSER(*), IFAIL
REAL (KIND=nag_wp)  BSCORE(NBEST), GAMMA, ACC(2), RUSER(*)
EXTERNAL  F

3  Description

Given Ω=xi:i,1im, a set of m unique features and a scoring mechanism f S  defined for all S Ω  then H05ABF is designed to find So1 Ω , So1 = p , an optimal subset of size p. Here So1 denotes the cardinality of So1, the number of elements in the set.
The definition of the optimal subset depends on the properties of the scoring mechanism, if
fSi fSj , for all ​ Sj Ω ​ and ​ Si Sj (1)
then the optimal subset is defined as one of the solutions to
maximize SΩ fS   subject to   S = p
else if
f Si fSj , for all ​ Sj Ω ​ and ​ Si Sj (2)
then the optimal subset is defined as one of the solutions to
minimize S Ω fS   subject to   S = p .
If neither of these properties hold then H05ABF cannot be used.
As well as returning the optimal subset, So1, H05ABF can return the best n solutions of size p. If Soi denotes the ith best subset, for i=1,2,,n-1, then the i+1th best subset is defined as the solution to either
maximize S Ω - Soj : j , 1ji fS   subject to   S = p
or
minimize S Ω - Soj : j,1ji fS   subject to   S = p
depending on the properties of f.
The solutions are found using a branch and bound method, where each node of the tree is a subset of Ω. Assuming that (1) holds then a particular node, defined by subset Si, can be trimmed from the tree if fSi < f^Son  where f^Son  is the nth highest score we have observed so far for a subset of size p, i.e., our current best guess of the score for the nth best subset. In addition, because of (1) we can also drop all nodes defined by any subset Sj where SjSi, thus avoiding the need to enumerate the whole tree. Similar short cuts can be taken if (2) holds. A full description of this branch and bound algorithm can be found in Ridout (1988).
Rather than calculate the score at a given node of the tree H05ABF utilizes the fast branch and bound algorithm of Somol et al. (2004), and attempts to estimate the score where possible. For each feature, xi, two values are stored, a count ci and μ^i, an estimate of the contribution of that feature. An initial value of zero is used for both ci and μ^i. At any stage of the algorithm where both f S  and f S - xi  have been calculated (as opposed to estimated), the estimated contribution of the feature xi is updated to
ciμ^i + f S - f S - xj ci+1
and ci is incremented by 1, therefore at each stage μ^i is the mean contribution of xi observed so far and ci is the number of observations used to calculate that mean.
As long as cik, for the user-supplied constant k, then rather than calculating f S - xi  this routine estimates it using f^ S - xi = f S - γ μ^i  or f^ S - γ μ^i  if f S  has been estimated, where γ is a user-supplied scaling factor. An estimated score is never used to trim a node or returned as the optimal score.
Setting k=0 in this routine will cause the algorithm to always calculate the scores, returning to the branch and bound algorithm of Ridout (1988). In most cases it is preferable to use the fast branch and bound algorithm, by setting k>0, unless the score function is iterative in nature, i.e., f S  must have been calculated before f S - xi  can be calculated.
H05ABF is a direct communication version of H05AAF.

4  References

Narendra P M and Fukunaga K (1977) A branch and bound algorithm for feature subset selection IEEE Transactions on Computers 9 917–922
Ridout M S (1988) Algorithm AS 233: An improved branch and bound algorithm for feature subset selection Journal of the Royal Statistics Society, Series C (Applied Statistics) (Volume 37) 1 139–147
Somol P, Pudil P and Kittler J (2004) Fast branch and bound algorithms for optimal feature selection IEEE Transactions on Pattern Analysis and Machine Intelligence (Volume 26) 7 900–912

5  Parameters

1:     MINCR – INTEGERInput
On entry: flag indicating whether the scoring function f is increasing or decreasing.
MINCR=1
fSi fSj , i.e., the subsets with the largest score will be selected.
MINCR=0
fSi fSj , i.e., the subsets with the smallest score will be selected.
For all SjΩ and SiSj.
Constraint: MINCR=0 or 1.
2:     M – INTEGERInput
On entry: m, the number of features in the full feature set.
Constraint: M2.
3:     IP – INTEGERInput
On entry: p, the number of features in the subset of interest.
Constraint: 1IPM.
4:     NBEST – INTEGERInput
On entry: n, the maximum number of best subsets required. The actual number of subsets returned is given by LA on final exit. If on final exit LANBEST then IFAIL=42 is returned.
Constraint: NBEST1.
5:     LA – INTEGEROutput
On exit: the number of best subsets returned.
6:     BSCORE(NBEST) – REAL (KIND=nag_wp) arrayOutput
On exit: holds the score for the LA best subsets returned in BZ.
7:     BZ(M-IP,NBEST) – INTEGER arrayOutput
On exit: the jth best subset is constructed by dropping the features specified in BZij, for i=1,2,,M-IP and j=1,2,,LA, from the set of all features, Ω. The score for the jth best subset is given in BSCOREj.
8:     F – SUBROUTINE, supplied by the user.External Procedure
F must evaluate the scoring function f.
The specification of F is:
SUBROUTINE F ( M, DROP, LZ, Z, LA, A, SCORE, IUSER, RUSER, INFO)
INTEGER  M, DROP, LZ, Z(LZ), LA, A(LA), IUSER(*), INFO
REAL (KIND=nag_wp)  SCORE(max(LA,1)), RUSER(*)
1:     M – INTEGERInput
On entry: m=Ω, the number of features in the full feature set.
2:     DROP – INTEGERInput
On entry: flag indicating whether the intermediate subsets should be constructed by dropping features from the full set (DROP=1) or adding features to the empty set (DROP=0). See SCORE for additional details.
3:     LZ – INTEGERInput
On entry: the number of features stored in Z.
4:     Z(LZ) – INTEGER arrayInput
On entry: Zi, for i=1,2,,LZ, contains the list of features which, along with those specified in A, define the subsets whose score is required. See SCORE for additional details.
5:     LA – INTEGERInput
On entry: if LA>0, the number of subsets for which a score must be returned.
If LA=0, the score for a single subset should be returned. See SCORE for additional details.
6:     A(LA) – INTEGER arrayInput
On entry: Aj, for j=1,2,,LA, contains the list of features which, along with those specified in Z, define the subsets whose score is required. See SCORE for additional details.
7:     SCORE(maxLA,1) – REAL (KIND=nag_wp) arrayOutput
On exit: the value fSj , for j=1,2,,LA, the score associated with the jth subset. Sj is constructed as follows:
DROP=1
Sj is constructed by dropping the features specified in the first LZ elements of Z and the single feature given in Aj from the full set of features, Ω. The subset will therefore contain M-LZ-1 features.
DROP=0
Sj is constructed by adding the features specified in the first LZ elements of Z and the single feature specified in Aj to the empty set, . The subset will therefore contain LZ+1 features.
In both cases the individual features are referenced by the integers 1 to M with 1 indicating the first feature, 2 the second, etc., for some arbitrary ordering of the features, chosen by you prior to calling H05ABF. For example, 1 might refer to the first variable in a particular set of data, 2 the second, etc..
If LA=0, the score for a single subset should be returned. This subset is constructed by adding or removing only those features specified in the first LZ elements of Z. If LZ=0, this subset will either be Ω or .
8:     IUSER(*) – INTEGER arrayUser Workspace
9:     RUSER(*) – REAL (KIND=nag_wp) arrayUser Workspace
F is called with the parameters IUSER and RUSER as supplied to H05ABF. You are free to use the arrays IUSER and RUSER to supply information to F as an alternative to using COMMON global variables.
10:   INFO – INTEGERInput/Output
On entry: INFO=0.
On exit: set INFO to a nonzero value if you wish H05ABF to terminate with IFAIL=82.
F must either be a module subprogram USEd by, or declared as EXTERNAL in, the (sub)program from which H05ABF is called. Parameters denoted as Input must not be changed by this procedure.
9:     MINCNT – INTEGERInput
On entry: k, the minimum number of times the effect of each feature, xi, must have been observed before f S - xi  is estimated from f S  as opposed to being calculated directly.
If k=0 then f S - xi  is never estimated. If MINCNT<0 then k is set to 1.
10:   GAMMA – REAL (KIND=nag_wp)Input
On entry: γ, the scaling factor used when estimating scores. If GAMMA<0 then γ is set to 1.
11:   ACC(2) – REAL (KIND=nag_wp) arrayInput
On entry: a measure of the accuracy of the scoring function, f.
Letting ai = ε1 fSi + ε2 , then when confirming whether the scoring function is strictly increasing or decreasing (as described in MINCR), or when assessing whether a node defined by subset Si can be trimmed, then any values in the range fSi ± ai  are treated as being numerically equivalent.
If 0ACC11 then ε1=ACC1, otherwise ε1=0.
If ACC20 then ε2=ACC2, otherwise ε2=0.
In most situations setting both ε1 and ε2 to zero should be sufficient. Using a nonzero value, when one is not required, can significantly increase the number of subsets that need to be evaluated.
12:   IUSER(*) – INTEGER arrayUser Workspace
13:   RUSER(*) – REAL (KIND=nag_wp) arrayUser Workspace
IUSER and RUSER are not used by H05ABF, but are passed directly to F and may be used to pass information to this routine as an alternative to using COMMON global variables.
14:   IFAIL – INTEGERInput/Output
On entry: IFAIL must be set to 0, -1​ or ​1. If you are unfamiliar with this parameter you should refer to Section 3.3 in the Essential Introduction for details.
For environments where it might be inappropriate to halt program execution when an error is detected, the value -1​ or ​1 is recommended. If the output of error messages is undesirable, then the value 1 is recommended. Otherwise, if you are not familiar with this parameter, the recommended value is 0. When the value -1​ or ​1 is used it is essential to test the value of IFAIL on exit.
On exit: IFAIL=0 unless the routine detects an error or a warning has been flagged (see Section 6).

6  Error Indicators and Warnings

If on entry IFAIL=0 or -1, explanatory error messages are output on the current error message unit (as defined by X04AAF).
Errors or warnings detected by the routine:
IFAIL=11
On entry, MINCR=value.
Constraint: MINCR=0 or 1.
IFAIL=21
On entry, M=value.
Constraint: M2.
IFAIL=31
On entry, IP=value and M=value.
Constraint: 1IPM. On entry, IP=value and M=value.
Constraint: 1IPM.
IFAIL=41
On entry, NBEST=value.
Constraint: NBEST1.
IFAIL=42
On entry, NBEST=value.
But only value best subsets could be calculated.
IFAIL=81
On exit from F, SCOREvalue=value, which is inconsistent with the score for the parent node. Score for the parent node is value. On exit from F, SCOREvalue=value, which is inconsistent with the score for the parent node. Score for the parent node is value.
IFAIL=82
A nonzero value for INFO has been returned: INFO=value.
IFAIL=-999
Dynamic memory allocation failed.

7  Accuracy

The subsets returned by H05ABF are guaranteed to be optimal up to the accuracy of the calculated scores.

8  Further Comments

The maximum number of unique subsets of size p from a set of m features is N= m! m-p!p! . The efficiency of the branch and bound algorithm implemented in H05ABF comes from evaluating subsets at internal nodes of the tree, that is subsets with more than p features, and where possible trimming branches of the tree based on the scores at these internal nodes as described in Narendra and Fukunaga (1977). Because of this it is possible, in some circumstances, for more than N subsets to be evaluated. This will tend to happen when most of the features have a similar effect on the subset score.
If multiple optimal subsets exist with the same score, and NBEST is too small to return them all, then the choice of which of these optimal subsets is returned is arbitrary.

9  Example

This example finds the three linear regression models, with five variables, that have the smallest residual sums of squares when fitted to a supplied dataset. The data used in this example was simulated.

9.1  Program Text

Program Text (h05abfe.f90)

9.2  Program Data

Program Data (h05abfe.d)

9.3  Program Results

Program Results (h05abfe.r)


H05ABF (PDF version)
H Chapter Contents
H Chapter Introduction
NAG Library Manual

© The Numerical Algorithms Group Ltd, Oxford, UK. 2012