0% found this document useful (0 votes)
45 views6 pages

Engineering Journal Missing Data Imputation Methods in Classification Contexts

Abstract— We examine different imputation methods that deal with missing data in classification contexts and compare the performance of the methods with an experiment study. We investigate the performance of the methods under the assumption that data are missing at random. We find that, as the number of missing holes in data increases, the imputation methods deteriorate and the misclassification rates of the imputation methods increase. We also examine the scenario where missing data are due to strategic behaviors of data providers. We find that imputation methods play an important role at deterring strategic behaviors of data providers and minimizing the misclassification rate.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
45 views6 pages

Engineering Journal Missing Data Imputation Methods in Classification Contexts

Abstract— We examine different imputation methods that deal with missing data in classification contexts and compare the performance of the methods with an experiment study. We investigate the performance of the methods under the assumption that data are missing at random. We find that, as the number of missing holes in data increases, the imputation methods deteriorate and the misclassification rates of the imputation methods increase. We also examine the scenario where missing data are due to strategic behaviors of data providers. We find that imputation methods play an important role at deterring strategic behaviors of data providers and minimizing the misclassification rate.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

International Journal of Engineering Research & Science (IJOER)

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

Missing Data Imputation Methods in Classification Contexts


Juheng Zhang
Department of Operations and Information Systems, University of Massachusetts Lowell, Lowell, MA

Abstract We examine different imputation methods that deal with missing data in classification contexts and compare the
performance of the methods with an experiment study. We investigate the performance of the methods under the assumption
that data are missing at random. We find that, as the number of missing holes in data increases, the imputation methods
deteriorate and the misclassification rates of the imputation methods increase. We also examine the scenario where missing
data are due to strategic behaviors of data providers. We find that imputation methods play an important role at deterring
strategic behaviors of data providers and minimizing the misclassification rate.
Keywords missing data, imputation method, classification.

I.

INTRODUCTION

Often in many empirical studies, data are missing due to various reasons. Missing data may be caused by negligence of data
collectors, poor experiment designs or procedures, or even purposely hiding behaviors of data providers. The two general
assumptions of missing data are: data missing at random and data missing strategically. Randomly missing data assumption
assumes that the missing data of an attribute are not related to the values themselves nor the values of other attributes. For
instance, in U.S. census data, a specific home address is missing, which is likely due to a random reason. As for strategically
missing data assumption, the data are missing due to strategic reasons. For instance, an insurance applicant can purposely
hide her/his smoking/drinking when apply for a health insurance in hope for a more likely result of approval. Another
example is limited information disclosure in financial markets [5, 6]. Certain companies strategically hide information from
investors. Missing data are a common problem in many research fields such as economics, marketing, health, statistics,
psychology, and education.
Missing data can lead to a number of problems [8]. The high level of statistical power requires a large amount of data. When
data are missing, sample size decreases dramatically if only observations with complete data are used. Empirical studies
found that if two percent of data are missing randomly in a data set, then eighteen percent of the total data can be lost when
observations having a missing value are removed. Missing data decreases statistical power.
In this study, we consider different imputation methods that either designed for randomly missing data or strategically
missing data. We compare the performance of the imputation methods in classification contexts under the assumption of data
missing at random. We also examine the imputation methods when data providers act strategically and data are hidden
intentionally. In the following section, we overview related research works and briefly discuss different imputation methods.

II.

LITERATURE REVIEW

In the statistics field, a few imputation methods such as the Average Method, the Similarity, and the Regression Method have
gained widespread acceptance. These methods normally consider attributes having continuous values. They have become
conventional methods for dealing with missing data, and we see adaption of these methods in different fields. We refer
readers to survey papers [4, 9] for detailed discussion on these conventional methods, although we discuss some of these
below. There are also some imputation techniques unique to classification problems. Several papers [2, 3, 7] summarize
different linear discriminate methods for handling missing data and compare the performance of these methods. The simplest
imputation method perhaps is the Average Method, also known as marginal mean imputation. The method was first
mentioned in the study [11]. For each missing value on a given variable, the Average Method finds the estimated value by
calculating the mean for those cases with observed data present on that variable. The Similarity Method finds an observation
that is the most similar to the record with a missing value as measured by the values not missing, and uses the actual value in
the most similar record to replace the missing value. Proponents of the Similarity Method argue that the method improves
accuracy since it uses realistic values, and that the method also preserves the shape of the distribution. The underlying
principal of the Similarity Method can be used for discrete variables, and this variant of the Similarity Method is called Hotdeck imputation, which has become popular in survey research. A data set is hot if it is currently being used for imputing a
score. The Hot-deck imputation replaces a missing value with actual scores of a similar case. If there are several equally
similar cases, then the method randomly chooses one of them. The Regression Method, also called the conditional mean
Page | 60

International Journal of Engineering Research & Science (IJOER)

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

imputation method, is another statistical imputation method. The Regression Method uses a regression equation to calculate
an estimate of the true value. Assume only one variable has missing values on some cases. Using the cases with complete
data on other variables, the method regresses on all of the other variables and then uses the regression equation to generate
the substitutes for the cases with missing data. The substitutes are predicted values for missing data. According to the study
[1], the Regression Method generates predicted values that preserve deviations from the mean and the shape of the
distribution. It does not attenuate correlations between variables as much as mean substitution.
Another category of imputation methods [12-14] assumes data are missing strategically by data providers who try to game
the decision makers decision rules. The imputation methods proposed in the studies [12, 13] include the D and DNeg
methods. These methods were designed for classification problems. The decision maker may use the D or DNeg method to
impute missing values and minimize misclassification rates when facing with strategic data providers. The DNeg method was
to thwart negative data providers from gaining a positive classification when they intentionally hide information. The D
method considered not only negative data providers but also positive ones. Using the D method, the decision maker can deter
negative data providers gaming behaviors and also incent positive data provider to reveal information. The D method is
more conservative than the DNeg method in a sense that it considers both positive and negative data providers while the
DNeg method is online for negative ones.

III.

EXPERIMENT DESIGN

We compare eight methods, Average, Regression, Similarity, D, DNeg, AvgNeg, RegNeg, and SimNeg. The Average,
Regression, Similarity, D, and DNeg are as what we discussed in the above section. The AvgNeg, RegNeg, and SimNeg are
the revised version of the original Average, Regression, and Similarity method respectively, in which only negative training
samples are used for imputing missing values. We first start with the case of randomly missing data, and then examine the
strategically missing data. The parameters of experimental design are listed in Table 1.

TABLE 1.
SUMMARY OF EXPERIMENTAL DESIGNS
Treatment
Replications
Dimensionality
Training set size
Testing set size
Randomly missing data percent
Data providers methods
Decision Makers methods
Outcomes

Parameter
30
3, 4, 5, 6, 7
20, 100, 200, 1000
20, 100, 200, 1000
1%, 2%, 3%, 4%, 5%
D, Average, Regression, Similarity, DNeg, AvgNeg, RegNeg, SimNeg
D, Average, Regression, Similarity, DNeg, AvgNeg, RegNeg, SimNeg
TotMisc, PosMisc, NegMisc, Notclassified, TotStrMisc, PosStrMisc,
NegStrMisc, StrPos, StrNeg

We use 30 replications for each case. The number of attributes ranges from 3 to 7. We use different training and testing set
sizes, 20, 100, 200, and 1000. In the randomly missing data case, we consider various percent of missing holes, 1%, 2%, 3%,
4%, and 5%. In the strategically missing data case, data providers may use one of the eight methods to hide information: D, Average,
Regression, Similarity, DNeg, AvgNeg, RegNeg, SimNeg. The decision maker chooses one of the eight methods to impute missing values.
We use different misclassification measurements. TotMisc is the misclassification over all data providers, PosMisc is the misclassification
rate over positive records, NegMisc is the misclassification rate over negative records, Notclassified is the rate that records not get
classified. To stabilize the variance of the rates of misclassification in statistical tests [10], we map the performance measures

to 2 arc sin (sqrt (misclassification rate)).

IV.

EMPIRICAL RESULTS

We first conduct an ANOVA analysis on the misclassification rate (TotMisc) for the randomly missing data case. We see
that all experiment factors are significant, as well as all interaction effect of all factors at 0.0001 confidence level. The
ANOVA analysis results are included in Table 2.

Page | 61

International Journal of Engineering Research & Science (IJOER)

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

TABLE 2.
TWO-WAY ANALYSIS FOR DEPENDENT VARIABLE: MISCLASSIFICATION IN RANDOM CASE

214

Sum of
Squares
3798.52

Mean
Square
17.75

Error

191785

3360.07

0.02

Corrected Total

191999

7158.59

Source

DF

Anova SS

Source

DF

F Value

Pr>F

R-Sqe

Model

1013.14

<.0001

0.53

Mean Square

F Value

Pr>F

98.43

24.61

1404.59

<.0001

4.71

1.57

89.63

<.0001

187.58

62.53

3568.79

<.0001

Ram

1546.32

386.58

22065.10

<.0001

Mp

1691.19

241.60

13789.90

<.0001

Coeff
Var
44.46

Root
MSE
0.13

Rate
Mean
0.30

Main effect

Two-way Interaction effect

12

2.93

0.24

13.94

<.0001

12

3.15

0.26

14.97

<.0001

n Ram

16

13.56

0.85

48.37

<.0001

n Mp

28

91.22

3.26

185.95

<.0001

2.13

0.24

13.49

<.0001

Ram

12

1.43

0.12

6.78

<.0001

Mp

21

47.44

2.26

128.94

<.0001

12

5.45

0.45

25.92

<.0001

21

5.57

0.27

15.13

<.0001

28

96.68

3.45

197.08

<.0001

Ram

Mp

Ram Mp

Table 2 shows that the percent of randomly missing data in datasets is a significant factor of misclassification rate. Next, we
study the impact of percent of randomly missing data on performance measurements in details. The results of the
misclassification rates for different percents of randomly missing data are provided in Table 3.

TABLE 3.
STATISTICS OF RANDOMLY MISSING HOLES
Missing Percent

TotMisc

PosMisc

NegMisc

NotClassified

1%

0.15957

0.21862

0.01514

0.00000

2%

0.24331

0.33372

0.02756

0.00006

3%

0.30635

0.42139

0.03891

0.00009

4%

0.36456

0.50299

0.04877

0.00026

5%

0.41469

0.57368

0.05807

0.00033

Page | 62

International Journal of Engineering Research & Science (IJOER)

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

As shown in Table 3, the misclassification rate over positive records is higher than that over negative records. In addition, all
of four misclassification rates, TotMisc, PosMisc, NegMisc, and NotClassified, increase as the percent increases. The results
in percentage format are provided in Table 4. In percent format, the misclassification rate is 0.64% when 1% data are missing
and increases to 4.24% when the percent of missing holes increases to 5%.

TABLE 4.
STATISTICS OF RANDOMLY MISSING HOLES IN PERCENT
Missing Percent

TotMisc

PosMisc

NegMisc

NotClassified

1%

0.640%

1.190%

0.010%

0.00%

2%

1.470%

2.760%

0.020%

0.00%

3%

2.330%

4.370%

0.040%

0.00%

4%

3.290%

6.190%

0.060%

0.00%

5%

4.240%

8.000%

0.080%

0.00%

We plot the trend of misclassification rate with the increase in the missing percent in Fig 1. The top line is for the
misclassification rate over positive records, and the TotMisc is the average over positive and negative misclassification rates.
The non-classified records stay as zero when the percent of missing holes increases.

In the random case, data are missing randomly, that is the information is not hidden strategically. A principal still can choose
one of eight methods to impute missing information. Next, we consider the case where data are missing strategically. We
simulate the case where agents select a method in determining which attributes to hide and simultaneously a decision maker
chooses from those eight methods to impute estimates for missing data of all agents. Similarly, we conduct an ANOVA test
to examine the effect of factors and interaction effect on the misclassification rates and provide the results in Table 5. The
hiding strategies of data providers is denoted as Ma . As shown in Table 5, the hiding strategies of data providers are
significant , and all other experiment factors are significant at the 0.0001 level.

Page | 63

International Journal of Engineering Research & Science (IJOER)

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

TABLE 5.
TWO-WAY ANALYSIS FOR DEPENDENT VARIABLE: MISCLASSIFICATION RATE IN THE STRATEGICALLY
MISSING CASE

DF

Sum of
Squares

Model

271

103371.3

381.4

Error

306928

7652.747

0.0249

Corrected Total

307199

111024.1

78.763

19.691

789.74

<.0001

66.17

22.057

884.62

<.0001

4.944

1.648

66.1

<.0001

Source

Mean
Square

F Value

Pr>F

R-Sq

Coeff
Var

Root
MSE

Rate
Mean

15298.5

<.0001

0.931

10.579

0.158

1.493

Main effect

Ma

2751.8

393.114

15766.6

<.0001

Mp

42912.157

6130.308

245868

<.0001

12

19.43

1.619

64.94

<.0001

12

1.015

0.085

3.39

<.0001

n Ma

28

430.728

15.383

616.97

<.0001

n Mp

28

1277.847

45.637

1830.37

<.0001

Two-way Interaction effect


Ma

0.56

0.062

2.5

0.0075

21

68.872

3.28

131.54

<.0001

Mp

21

156.013

7.429

297.96

<.0001

Ma

21

1.672

0.08

3.19

<.0001

21

13.682

0.652

26.13

<.0001

49

25245.6

515.218

20663.8

<.0001

Mp
Ma Mp

We conduct Tukeys range tests for the decision makers methods. The summary of these test results can be found in Tables
6. Table 6 shows that when the decision maker uses the D or DNeg method, NegMisc is the lowest and PosMisc is the
highest. More specifically, we see that NegMisc is 0 for the D method and 0.013 for the DNeg method (or 0.004% in terms
of actual rate before the mapping), which are lower than 0.402 (or 3.98% ) for Similarity, 0.324 (or 2.6%) for RegNeg, 0.322
(or 2.58%) for Regression, 0.283 (or 1.99%) for Average, 0.06 (or 0.09%) for AvgNeg, and 0.031 (or 0.02%) for SimNeg.
The rate of positive or negative agents who act strategically is the same regardless of a decision makers method. Therefore,
if a decision maker is more negative risk averse, her best strategy should be the D method.

TABLE 6.
THE COMPARISON RESULTS OF TUKEY'S RANGE TEST
TotMisc

PosMisc

NegMisc

NotClassified

Group

Mean

Methods

Group

Mean

Methods

Group

Mean

Methods

Group

Mean

Methods

2.205

2.192

0.402

1.902

2.086

1.893

0.324

1.902

2.065

1.629

0.322

1.902

2.062

1.283

0.283

1.902

1.128

0.783

0.06

0.992

0.185

0.031

0.895

0.167

FE

0.013

0.822

0.158

5
Page | 64

International Journal of Engineering Research & Science (IJOER)

V.

ISSN: [2395-6992]

[Vol-2, Issue-4 April- 2016]

CONCLUSION

We compare eight different imputation methods in the case where data are missing at random and in the case where data are
missing strategically. We find that as the percent of missing data increases, the performance of all the eight imputation
methods decreases. When data are missing strategically, the D method or DNeg method gives the lowest misclassification
rate.

REFERENCES
[1] Allison, P.D. Missing data. Woburn, MA, USA: Sage Publications Inc., 2001.
[2] Chan, L.S., and Dunn, O.J. The treatment of missing values in discriminant analysis-1. The sampling experiment. Journal of the
American Statistical Association, 67, 338 (1972), 473-477.
[3] Chan, L.S., Gilman, J.A., and Dunn, O.J. Alternative approaches to missing values in discriminant analysis. Journal of the American
Statistical Association, 71, 356 (1976), 842-844.
[4] Donders, A.R.T., van der Heijden, G., Stijnen, T., and Moons, K.G.M. Review: A gentle introduction to imputation of missing values.
Journal of Clinical Epidemiology, 59, 10 (2006), 1087-1091.
[5] Healy, P.M., and Palepu, K.G. The challenges of investor communication the case of cuc international, inc. Journal of financial
economics, 38, 2 (1995), 111-140.
[6] Hirshleifer, D., and Teoh, S.H. Limited attention, information disclosure, and financial reporting. Journal of accounting and
economics, 36, 1-3 (2003), 337-386.
[7] Jackson, E.C. Missing values in linear multiple discriminant analysis. Biometrics, 24, 4 (1968), 835-844.
[8] Roth, P.L. Missing data: A conceptual review for applied psychologists. Personnel Psychology, 47, 3 (1994), 537-560.
[9] Schafer, J.L., and Graham, J.W. Missing data: Our view of the state of the art. Psychological Methods, 7, 2 (2002), 147-177.
[10] Stam, A., and Joachimsthaler, E.A. Solving the classification problem in discriminant analysis via linear and nonlinear programming
methods. Decision Sciences, 20, 2 (1989), 285-293.
[11] Wilks, S.S. Moments and distributions of estimates of population parameters from fragmentary samples. The Annals of Mathematical
Statistics, 3, 3 (1932), 163-195.
[12] Zhang, J. Linear discrimination with strategic missing values. Information Systems and Operations Management, Gainesville, FL,
USA: University of Florida, 2011.
[13] Zhang, J., Aytug, H., and Koehler, G.J. Discriminant analysis with strategically manipulated data. Information Systems Research, 25,
3 (2014), 654-662.
[14] Zhang, J., Liu, X., and Li, X. Support vector regression for handling strategically hidden data. Lowell, U.o.M., 2015

Page | 65

You might also like