"""Data check that checks if the target data contains missing or invalid values."""
import woodwork as ww
from evalml.data_checks import (
DataCheck,
DataCheckActionCode,
DataCheckActionOption,
DataCheckError,
DataCheckMessageCode,
DataCheckWarning,
DCAOParameterType,
)
from evalml.objectives import get_objective
from evalml.problem_types import (
handle_problem_types,
is_binary,
is_multiclass,
is_regression,
)
from evalml.utils.woodwork_utils import infer_feature_types, numeric_and_boolean_ww
[docs]class InvalidTargetDataCheck(DataCheck):
"""Check if the target data is considered invalid.
Target data is considered invalid if:
- Target is None.
- Target has NaN or None values.
- Target is of an unsupported Woodwork logical type.
- Target and features have different lengths or indices.
- Target does not have enough instances of a class in a classification problem.
- Target does not contain numeric data for regression problems.
Args:
problem_type (str or ProblemTypes): The specific problem type to data check for.
e.g. 'binary', 'multiclass', 'regression, 'time series regression'
objective (str or ObjectiveBase): Name or instance of the objective class.
n_unique (int): Number of unique target values to store when problem type is binary and target
incorrectly has more than 2 unique values. Non-negative integer. If None, stores all unique values. Defaults to 100.
null_strategy (str): The type of action option that should be returned if the target is partially null.
The options are `impute` and `drop` (default).
`impute` - Will return a `DataCheckActionOption` for imputing the target column.
`drop` - Will return a `DataCheckActionOption` for dropping the null rows in the target column.
"""
multiclass_continuous_threshold = 0.05
def __init__(self, problem_type, objective, n_unique=100, null_strategy="drop"):
self.problem_type = handle_problem_types(problem_type)
self.objective = get_objective(objective)
if n_unique is not None and n_unique <= 0:
raise ValueError("`n_unique` must be a non-negative integer value.")
self.n_unique = n_unique
if null_strategy is None or null_strategy.lower() not in ["impute", "drop"]:
raise ValueError(
"The acceptable values for 'null_strategy' are 'impute' and 'drop'.",
)
self.null_strategy = null_strategy
[docs] def validate(self, X, y):
"""Check if the target data is considered invalid. If the input features argument is not None, it will be used to check that the target and features have the same dimensions and indices.
Target data is considered invalid if:
- Target is None.
- Target has NaN or None values.
- Target is of an unsupported Woodwork logical type.
- Target and features have different lengths or indices.
- Target does not have enough instances of a class in a classification problem.
- Target does not contain numeric data for regression problems.
Args:
X (pd.DataFrame, np.ndarray): Features. If not None, will be used to check that the target and features have the same dimensions and indices.
y (pd.Series, np.ndarray): Target data to check for invalid values.
Returns:
dict (DataCheckError): List with DataCheckErrors if any invalid values are found in the target data.
Examples:
>>> import pandas as pd
Target values must be integers, doubles, or booleans.
>>> X = pd.DataFrame({"col": [1, 2, 3, 1]})
>>> y = pd.Series(["cat_1", "cat_2", "cat_1", "cat_2"])
>>> target_check = InvalidTargetDataCheck("regression", "R2", null_strategy="impute")
>>> assert target_check.validate(X, y) == [
... {
... "message": "Target is unsupported Unknown type. Valid Woodwork logical types include: integer, double, boolean, age, age_fractional, integer_nullable, boolean_nullable, age_nullable",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "error",
... "details": {"columns": None, "rows": None, "unsupported_type": "unknown"},
... "code": "TARGET_UNSUPPORTED_TYPE",
... "action_options": []
... },
... {
... "message": "Target data type should be numeric for regression type problems.",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "error",
... "details": {"columns": None, "rows": None},
... "code": "TARGET_UNSUPPORTED_TYPE_REGRESSION",
... "action_options": []
... }
... ]
The target cannot have null values.
>>> y = pd.Series([None, pd.NA, pd.NaT, None])
>>> assert target_check.validate(X, y) == [
... {
... "message": "Target is either empty or fully null.",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "error",
... "details": {"columns": None, "rows": None},
... "code": "TARGET_IS_EMPTY_OR_FULLY_NULL",
... "action_options": []
... }
... ]
...
...
>>> y = pd.Series([1, None, 3, None])
>>> assert target_check.validate(None, y) == [
... {
... "message": "2 row(s) (50.0%) of target values are null",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "error",
... "details": {
... "columns": None,
... "rows": [1, 3],
... "num_null_rows": 2,
... "pct_null_rows": 50.0
... },
... "code": "TARGET_HAS_NULL",
... "action_options": [
... {
... "code": "IMPUTE_COL",
... "data_check_name": "InvalidTargetDataCheck",
... "parameters": {
... "impute_strategy": {
... "parameter_type": "global",
... "type": "category",
... "categories": ["mean", "most_frequent"],
... "default_value": "mean"
... }
... },
... "metadata": {"columns": None, "rows": None, "is_target": True},
... }
... ],
... }
... ]
If the target values don't match the problem type passed, an error will be raised.
In this instance, only two values exist in the target column, but multiclass has been passed as the problem type.
>>> X = pd.DataFrame([i for i in range(50)])
>>> y = pd.Series([i%2 for i in range(50)])
>>> target_check = InvalidTargetDataCheck("multiclass", "Log Loss Multiclass")
>>> assert target_check.validate(X, y) == [
... {
... "message": "Target has two or less classes, which is too few for multiclass problems. Consider changing to binary.",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "error",
... "details": {"columns": None, "rows": None, "num_classes": 2},
... "code": "TARGET_MULTICLASS_NOT_ENOUGH_CLASSES",
... "action_options": []
... }
... ]
If the length of X and y differ, a warning will be raised. A warning will also be raised for indices that don"t match.
>>> target_check = InvalidTargetDataCheck("regression", "R2")
>>> X = pd.DataFrame([i for i in range(5)])
>>> y = pd.Series([1, 2, 4, 3], index=[1, 2, 4, 3])
>>> assert target_check.validate(X, y) == [
... {
... "message": "Input target and features have different lengths",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "warning",
... "details": {"columns": None, "rows": None, "features_length": 5, "target_length": 4},
... "code": "MISMATCHED_LENGTHS",
... "action_options": []
... },
... {
... "message": "Input target and features have mismatched indices. Details will include the first 10 mismatched indices.",
... "data_check_name": "InvalidTargetDataCheck",
... "level": "warning",
... "details": {
... "columns": None,
... "rows": None,
... "indices_not_in_features": [],
... "indices_not_in_target": [0]
... },
... "code": "MISMATCHED_INDICES",
... "action_options": []
... }
... ]
"""
messages = []
if y is None:
messages.append(
DataCheckError(
message="Target is None",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_IS_NONE,
details={},
).to_dict(),
)
return messages
y = infer_feature_types(y)
messages = self._check_target_has_nan(y, messages)
if any(error["code"] == "TARGET_IS_EMPTY_OR_FULLY_NULL" for error in messages):
# If our target is empty or fully null, no need to check for other invalid targets, return immediately.
return messages
messages = self._check_target_is_unsupported_type(y, messages)
messages = self._check_regression_target(y, messages)
messages = self._check_classification_target(y, messages)
messages = self._check_for_non_positive_target(y, messages)
messages = self._check_target_and_features_compatible(X, y, messages)
return messages
def _check_target_is_unsupported_type(self, y, messages):
is_supported_type = y.ww.logical_type.type_string in numeric_and_boolean_ww + [
ww.logical_types.Categorical.type_string,
]
if not is_supported_type:
messages.append(
DataCheckError(
message="Target is unsupported {} type. Valid Woodwork logical types include: {}".format(
type(y.ww.logical_type),
", ".join([ltype for ltype in numeric_and_boolean_ww]),
),
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_UNSUPPORTED_TYPE,
details={"unsupported_type": y.ww.logical_type.type_string},
).to_dict(),
)
return messages
def _check_target_has_nan(self, y, messages):
null_rows = y.isnull()
if null_rows.all():
messages.append(
DataCheckError(
message="Target is either empty or fully null.",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_IS_EMPTY_OR_FULLY_NULL,
details={},
).to_dict(),
)
return messages
elif null_rows.any():
num_null_rows = null_rows.sum()
pct_null_rows = null_rows.mean() * 100
rows_to_drop = null_rows.loc[null_rows].index.tolist()
action_options = []
impute_action_option = DataCheckActionOption(
DataCheckActionCode.IMPUTE_COL,
data_check_name=self.name,
parameters={
"impute_strategy": {
"parameter_type": DCAOParameterType.GLOBAL,
"type": "category",
"categories": (
["mean", "most_frequent"]
if is_regression(self.problem_type)
else ["most_frequent"]
),
"default_value": (
"mean"
if is_regression(self.problem_type)
else "most_frequent"
),
},
},
metadata={"is_target": True},
)
drop_action_option = DataCheckActionOption(
DataCheckActionCode.DROP_ROWS,
data_check_name=self.name,
metadata={"is_target": True, "rows": rows_to_drop},
)
if self.null_strategy.lower() == "impute":
action_options.append(impute_action_option)
elif self.null_strategy.lower() == "drop":
action_options.append(drop_action_option)
messages.append(
DataCheckError(
message="{} row(s) ({}%) of target values are null".format(
num_null_rows,
pct_null_rows,
),
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_HAS_NULL,
details={
"num_null_rows": num_null_rows,
"pct_null_rows": pct_null_rows,
"rows": rows_to_drop,
},
action_options=action_options,
).to_dict(),
)
return messages
def _check_target_and_features_compatible(self, X, y, messages):
if X is not None:
X = infer_feature_types(X)
X_index = list(X.index)
y_index = list(y.index)
X_length = len(X_index)
y_length = len(y_index)
if X_length != y_length:
messages.append(
DataCheckWarning(
message="Input target and features have different lengths",
data_check_name=self.name,
message_code=DataCheckMessageCode.MISMATCHED_LENGTHS,
details={
"features_length": X_length,
"target_length": y_length,
},
).to_dict(),
)
if X_index != y_index:
if set(X_index) == set(y_index):
messages.append(
DataCheckWarning(
message="Input target and features have mismatched indices order.",
data_check_name=self.name,
message_code=DataCheckMessageCode.MISMATCHED_INDICES_ORDER,
details={},
).to_dict(),
)
else:
index_diff_not_in_X = list(set(y_index) - set(X_index))[:10]
index_diff_not_in_y = list(set(X_index) - set(y_index))[:10]
messages.append(
DataCheckWarning(
message="Input target and features have mismatched indices. Details will include the first 10 mismatched indices.",
data_check_name=self.name,
message_code=DataCheckMessageCode.MISMATCHED_INDICES,
details={
"indices_not_in_features": index_diff_not_in_X,
"indices_not_in_target": index_diff_not_in_y,
},
).to_dict(),
)
return messages
def _check_for_non_positive_target(self, y, messages):
any_neg = (
not (y > 0).all()
if y.ww.logical_type.type_string
in [
ww.logical_types.Integer.type_string,
ww.logical_types.Double.type_string,
]
else None
)
if any_neg and self.objective.positive_only:
details = {
"Count of offending values": sum(
val <= 0 for val in y.values.flatten()
),
}
messages.append(
DataCheckError(
message=f"Target has non-positive values which is not supported for {self.objective.name}",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_INCOMPATIBLE_OBJECTIVE,
details=details,
).to_dict(),
)
return messages
def _check_regression_target(self, y, messages):
if is_regression(self.problem_type) and "numeric" not in y.ww.semantic_tags:
messages.append(
DataCheckError(
message="Target data type should be numeric for regression type problems.",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_UNSUPPORTED_TYPE_REGRESSION,
details={},
).to_dict(),
)
return messages
def _check_classification_target(self, y, messages):
value_counts = y.value_counts()
unique_values = value_counts.index.tolist()
if is_binary(self.problem_type) and len(value_counts) != 2:
if self.n_unique is None:
details = {"target_values": unique_values}
else:
details = {
"target_values": unique_values[
: min(self.n_unique, len(unique_values))
],
}
messages.append(
DataCheckError(
message="Binary class targets require exactly two unique values.",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_BINARY_NOT_TWO_UNIQUE_VALUES,
details=details,
).to_dict(),
)
elif is_multiclass(self.problem_type):
if value_counts.min() <= 1:
least_populated = value_counts[value_counts <= 1]
details = {
"least_populated_class_labels": sorted(
least_populated.index.tolist(),
),
}
messages.append(
DataCheckError(
message="Target does not have at least two instances per class which is required for multiclass classification",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_MULTICLASS_NOT_TWO_EXAMPLES_PER_CLASS,
details=details,
).to_dict(),
)
if len(unique_values) <= 2:
details = {"num_classes": len(unique_values)}
messages.append(
DataCheckError(
message="Target has two or less classes, which is too few for multiclass problems. Consider changing to binary.",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_MULTICLASS_NOT_ENOUGH_CLASSES,
details=details,
).to_dict(),
)
num_class_to_num_value_ratio = len(unique_values) / len(y)
if num_class_to_num_value_ratio >= self.multiclass_continuous_threshold:
details = {"class_to_value_ratio": num_class_to_num_value_ratio}
messages.append(
DataCheckWarning(
message="Target has a large number of unique values, could be regression type problem.",
data_check_name=self.name,
message_code=DataCheckMessageCode.TARGET_MULTICLASS_HIGH_UNIQUE_CLASS,
details=details,
).to_dict(),
)
return messages