Anda di halaman 1dari 3

Discard the examples that contain the missing values

convert the missing values into a new value (encode them into a new numerical value)
Such simplistic method was shown to lead to serious inference problems. However, these method are
practical only when the data contain relatively small number of examples with missing values and when
analysis of the complete examples will not lead to serious bias during the inference.

Imputation methods
A fundamental advantage of this approach is that the missing data treatment is independent of the
learning algorithm used.
Three different mechanism of missing data induction.

1. Missing completely at random (MCAR), when the distribution of an example having a missing value
for an attribute does not depend on either the observed data or the missing data.

2. Missing at random (MAR), when the distribution of an example having a missing value for an
attribute depends on the observed data, but does not depend on the missing data.

3. Not missing at random (NMAR), when the distribution of an example having a missing value for an
attribute depends on the missing values.
Imputation Methods
Global Average Value for Numerical Attributes (MC)
This method is very simple: for nominal attributes, the MV is replaced with the most common attribute
value, and numerical values are replaced with the average of all values of the corresponding attribute.
Concept Average Value for Numerical Attributes (CMC)
Replace the MV by the most repeated one if nominal or the mean value if numerical, but considering
only the instances with same class as the reference instance.
K-Nearest Neighbor (KNNI)
Using this instance-based algorithm, every time a MV found in a current instance, we compute the k
nearest neighbours and impute a value from them. For nominal values, the most common value among
all neighbours is taken, and for numerical values we will use the average value. Indeed, we need to
define a proximity measure between instances. We have chosen euclidean distance (it is a case of a Lp
norm distance), which is usually used
K-means Clustering Imputation (KMI)
Data objects that belong to the same cluster are taken as near- est neighbours of each other, and we
apply a nearest neighbour algorithm to replace missing data, in a similar way than k-Nearest Neighbour
Imputation.
Support Vector Machines Imputation (SVMI)
SVM regression based algorithm to fill in missing data, i.e. set the decision attributes (output or classes)
as the condition attributes (input attributes) and the con- dition attributes as the decision attributes, so
we can use SVM regression to predict the missing condition attribute values. In order to do that, first we
select the examples in which there are no missing attribute values. In the next step we set one of the
condition attributes (input attribute), some of those values are missing, as the decision attribute (output
attribute), and the decision attributes as the condition attributes by contraries. Finally, we use SVM
regression to predict the decision attribute values.
Imputation with Fuzzy K-means Clustering (FKMI)
FKMI, SVMI and EC methods for the approximate models.
In fuzzy clustering, each data object xi has a membership function which describes the degree which this
data object belongs to a certain cluster vk. In the process of updating membership functions and
centroids, we only take into account only complete attributes. In this process, we cannot assign the data
object to a concrete cluster represented by a cluster centroid (as done in the basic K-mean clustering
algorithm), because each data object belongs to all K clusters with different membership degrees. We
replace non-reference attributes for each incomplete data object xi based on the information about
membership degrees and the values of cluster centroids.
Event Covering (EC)
Based on the work of Wong et al., a mixed-mode probability model is approximated by a discrete one.
First, they discretize the continuous components using a minimum loss of information criterion. Treating
a mixed-mode feature n-tuple as a discrete-valued one, the authors propose a new statistical approach
for synthesis of knowledge based on cluster analysis. As main advantage, this method does not require
neither scale normalization nor ordering of discrete values. By synthesis of the data into statistical
knowledge, they refer to the following processes: 1) synthesize and detect from data inherent patterns
which indicate statistical interdependency; 2) group the given data into inherent clusters based on these
detected interdependency; and 3) interpret the underlying patterns for each clusters identified. The
method of synthesis is based on author's event {covering approach. With the developed inference
method, we are able to estimate the MVs in the data.

Anda mungkin juga menyukai