Predicting Glaucomatous Progression using Trees and Random Forests

Report 2 Downloads 32 Views
PPIG

Classification Tree

Random forests

PPIG revisited

Predicting Glaucomatous Progression using Trees and Random Forests Richard A. Levine San Diego State University Joint work with Lucie Sharpsten, Juanjuan Fan, Xiaogang Su, & Shaban Demirel

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Discussion

Perimetry and Psychophysics in Glaucoma (PPIG) Study

• 166 subjects with moderate- to high-risk ocular hypertension

• 332 eyes, subjects average age 58.1 (11.0 std), range 33 to 87

years

• 95 females, 71 males recruited in Portland, OR metro area • Median follow-up 6+ years, maximum 9+ years

• Outcome: progressive glaucomatous optic neuropathy (pGON) • 26 subjects pGON both eyes; 41 pGON in one eye • Goal: predict pGON using BL & FU data

PPIG

Classification Tree

Random forests

PPIG revisited

Humphrey Visual Field Analyzer: SAP

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Clinical factors used in data analysis Both baseline and follow-up: • 52 test point thresholds • mean deviation (MD)

• pattern standard deviation (PSD)

Baseline only: • intraocular pressure (IOP in mmHg)

• central corneal thickness (CCT in microns) • gender indicator • age (in years)

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decision tree TP44 ≥ 30.17

TP1 < 30.95

13; 0%

TP1 < 32.12

TP32 ≥ 32.29

37; 22%

19; 42%

IOP < 15.5

7; 0%

TP53 ≥ 30.85

7; 14%

30; 23%

12; 100%

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

CART (Breiman et al., 1984)

• Regression problem: Y outcome/output, X covariates/input

• Why tree methods?

Fewer assumptions; interpretation; prognostic rules

• Partitioning covariate space • Three steps to the CART algorithm: 1. grow a large tree 2. prune to a nested sequence of subtrees 3. select the best-sized tree

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decisions trees for correlated binary data (Sharpsten et al., 2013)

Growing a large tree T0 • Marginal model, i = 1, . . . , n; k = 1, 2; j = 1, . . . , p

logit(pik ) = β0 + β1 I (Xikj ≤ c) I (Xikj ≤ c) or I (Xikj ∈ A) induces a split

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decisions trees for correlated binary data (Sharpsten et al., 2013)

Growing a large tree T0 • Marginal model, i = 1, . . . , n; k = 1, 2; j = 1, . . . , p

logit(pik ) = β0 + β1 I (Xikj ≤ c) I (Xikj ≤ c) or I (Xikj ∈ A) induces a split

• Splitting criteria: maximum robust Z from GEE

(for testing H0 : β1 = 0.)

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Summarizing longitudinal measurements (Sharpsten et al., 2013)

• PLRA (Vesti et al., 2003, IOVS);

VFI (Leung et al., 2010, IOVS)

• Generalized least squares: visits l = 1, . . . , vi

Xl = α + βtl + ζl • Time windows (Gardiner et al., 2013, IOVS)

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Discussion

Variable importance via extremely randomized trees Algorithm idea from Geurts et al. (2006, ML) • Generate bootstrap samples Lb , b = 1, . . . , B • Grow a tree Tb using Lb

(m randomly selected inputs at each split; 1-3 random cut-points per input; no pruning)

• Compute G (Tb ) using out-of-bag sample L − Lb • For all predictors Zj , j = 1, . . . , p • Permute the values of Zj in L − Lb • Compute Gj (Tb ) using permuted L − Lb

• Compute variable importance

Vj =

1 B

P

b {G (Tb )

− Gj (Tb )}/G (Tb )

PPIG

Classification Tree

Random forests

PPIG revisited

Prognostic groups

Group I II

Definition Less likely More likely

Number of eyes 166 166

Number (%) of progressed eyes 10 (6.4%) 83 (50.0%)

Robust Z II 58.7 –

Discussion

Classification Tree

Random forests

PPIG revisited

0.6 0.4 0.2

True positive rate

0.8

1.0

ROC comparisons

Proposed Method Best Tree via GEE Robust Z Demirel’s average tree 0.0

PPIG

0.0

0.2

0.4

0.6

False positive rate

0.8

1.0

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Predicting pGON: Test-point importance

Discussion

PPIG

Classification Tree

Random forests

Current and future work • Joint models

• Bayesian trees

• Random forest variable importance bias

PPIG revisited

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Discussion

Relevant decision tree literature

• Segal (1992, JASA): longitudinal quantitative responses

extending CART least squares split function

• Zhang (1998, JASA): multiple binary data using generalized

entropy criterion

• Lee (2005, DMKD): multiple binary data GEE residual-based

approach

• Fan et al. (2006, JASA): multivariate survival data

PPIG

Classification Tree

Random forests

PPIG revisited

Decision tree TP44 ≥ 30.17

TP1 < 30.95

13; 0%

TP1 < 32.12

TP32 ≥ 32.29

37; 22%

19; 42%

IOP < 15.5

7; 0%

TP53 ≥ 30.85

7; 14%

30; 23%

12; 100%

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decisions trees for correlated binary data Pruning algorithm • T0 fully grown tree

• Solve for α at each internal node h

Gα (Th ) = G (Th ) − α|Thi | = 0, P where G (T ) = h∈T i G (h). • Prune tree at the internal node h with smallest α. • Repeat to obtain sequences

α1 , α2 , . . . , αM T0  T1  · · ·  TM

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decisions trees for correlated binary data Pruning algorithm • T0 fully grown tree

• Solve for α at each internal node h

Gα (Th ) = G (Th ) − α|Thi | = 0, P where G (T ) = h∈T i G (h). • Prune tree at the internal node h with smallest α. • Repeat to obtain sequences

α1 , α2 , . . . , αM T0  T1  · · ·  TM Theorem: For any α ∈ [αm , αm+1 ], Tm is best subtree.

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Decisions trees for correlated binary data Selecting the best-sized subtree • For fixed α in the range of 2 ≤ α ≤ 4,

ˆα (Tm ) max G

1≤m≤M

Discussion

PPIG

Classification Tree

Random forests

PPIG revisited

Discussion

Decisions trees for correlated binary data Selecting the best-sized subtree • For fixed α in the range of 2 ≤ α ≤ 4,

ˆα (Tm ) max G

1≤m≤M

• Bootstrap bias corrected estimate:

0 = bootstrap sample Lb , b = 1, . . . , B; αm



αm αm+1

0 ˆ (Tm ) = G (L; L, T (αm G )) −

(1/B)

B X  b=1

0 0 G (Lb ; Lb , Tb (αm )) − G (L; Lb , Tb (αm ))

PPIG

Classification Tree

Random forests

PPIG revisited

Discussion

Best tree for predicting pGON

TP44 ≤ 31.46

! " ! " #$ " ! !"

3 39

TP4 ≤ 22.29

+ + +

I

)) )) ))

, , ,#$

!" ** ** II **

TP8 ≤ 28.52

14 23

TP6 ≤ 30.98

) ) )) ))

TP27 ≤ 27.76

%% % %

& &

TP12 ≤ 31.47

!"

I

TP13 ≤ 26.76

&&

TP32 ≤ 30.58

! " ! " #$ " !

1 14

TP44 ≤ 30.53

## $ $$#$ #$ # !"

!"

6 16

20 37

II

II

TP27 ≤ 16.04

** ** **

'' ( (( #$ #$ ' !" !"

2 43

8 17

I

II

## $ $$#$ #$ # !"

!"

15 33

13 28

II

II

'' ( (( #$ #$ ' !" !"

4 70

7 12

I

II

PPIG

Classification Tree

Random forests

PPIG follow-up data

PPIG revisited

Discussion