nvtabular.ops.FillMedian#
- class nvtabular.ops.FillMedian(add_binary_cols=False)[source]#
Bases:
StatOperator
This operation replaces missing values with the median value for the column.
Example usage:
# Use FillMedian in a workflow for continuous columns cont_features = ['cont1', 'cont2', 'cont3'] >> ops.FillMedian() processor = nvtabular.Workflow(cont_features)
- Parameters:
add_binary_cols (boolean, default False) – When True, adds binary columns that indicate whether cells in each column were filled
Methods
__init__
([add_binary_cols])clear
()column_mapping
(col_selector)compute_column_schema
(col_name, input_schema)compute_input_schema
(root_schema, ...)Given the schemas coming from upstream sources and a column selector for the input columns, returns a set of schemas for the input columns this operator will use
compute_output_schema
(input_schema, col_selector)Given a set of schemas and a column selector for the input columns, returns a set of schemas for the transformed columns this operator will produce
compute_selector
(input_schema, selector[, ...])Provides a hook method for sub-classes to override to implement custom column selection logic.
create_node
(selector)export
(path, input_schema, output_schema, ...)Export the class object as a config and all related files to the user defined path.
fit
(col_selector, ddf)Calculate statistics for this operator, and return a dask future to these statistics, which will be computed by the workflow.
fit_finalize
(dask_stats)Finalize statistics calculation - the workflow calls this function with the computed statistics from the 'fit' object'
inference_initialize
(col_selector, model_config)Configures this operator for use in inference.
load_artifacts
([artifact_path])Load artifacts from disk required for operator function.
output_column_names
(col_selector)Given a set of columns names returns the names of the transformed columns this operator will produce
save_artifacts
([artifact_path])Save artifacts required to be reload operator state from disk
set_storage_path
(new_path[, copy])Certain stat operators need external storage - for instance Categorify writes out parquet files containing the categorical mapping.
transform
(col_selector, df)Transform the dataframe by applying this operator to the set of input columns
validate_schemas
(parents_schema, ...[, ...])Provides a hook method that sub-classes can override to implement schema validation logic.
Attributes
dependencies
Defines an optional list of column dependencies for this operator.
dynamic_dtypes
export_name
Provides a clear common english identifier for this operator.
is_subgraph
label
output_dtype
output_properties
output_tags
supported_formats
supports
Returns what kind of data representation this operator supports
- transform(col_selector: ColumnSelector, df: DataFrame) DataFrame [source]#
Transform the dataframe by applying this operator to the set of input columns
- fit(col_selector: ColumnSelector, ddf: DataFrame)[source]#
Calculate statistics for this operator, and return a dask future to these statistics, which will be computed by the workflow.