View source on GitHub |
DenseColumn
that converts from sparse, categorical input.
tf.feature_column.embedding_column(
categorical_column, dimension, combiner='mean', initializer=None,
ckpt_to_load_from=None, tensor_name_in_ckpt=None, max_norm=None, trainable=True
)
Use this when your inputs are sparse, but you want to convert them to a dense representation (e.g., to feed to a DNN).
Inputs must be a CategoricalColumn
created by any of the
categorical_column_*
function. Here is an example of using
embedding_column
with DNNClassifier
:
video_id = categorical_column_with_identity(
key='video_id', num_buckets=1000000, default_value=0)
columns = [embedding_column(video_id, 9),...]
estimator = tf.estimator.DNNClassifier(feature_columns=columns, ...)
label_column = ...
def input_fn():
features = tf.io.parse_example(
..., features=make_parse_example_spec(columns + [label_column]))
labels = features.pop(label_column.name)
return features, labels
estimator.train(input_fn=input_fn, steps=100)
Here is an example using embedding_column
with model_fn:
def model_fn(features, ...):
video_id = categorical_column_with_identity(
key='video_id', num_buckets=1000000, default_value=0)
columns = [embedding_column(video_id, 9),...]
dense_tensor = input_layer(features, columns)
# Form DNN layers, calculate loss, and return EstimatorSpec.
...
categorical_column
: A CategoricalColumn
created by a
categorical_column_with_*
function. This column produces the sparse IDs
that are inputs to the embedding lookup.dimension
: An integer specifying dimension of the embedding, must be > 0.combiner
: A string specifying how to reduce if there are multiple entries in
a single row. Currently 'mean', 'sqrtn' and 'sum' are supported, with
'mean' the default. 'sqrtn' often achieves good accuracy, in particular
with bag-of-words columns. Each of this can be thought as example level
normalizations on the column. For more information, see
tf.embedding_lookup_sparse
.initializer
: A variable initializer function to be used in embedding
variable initialization. If not specified, defaults to
truncated_normal_initializer
with mean 0.0
and
standard deviation 1/sqrt(dimension)
.ckpt_to_load_from
: String representing checkpoint name/pattern from which to
restore column weights. Required if tensor_name_in_ckpt
is not None
.tensor_name_in_ckpt
: Name of the Tensor
in ckpt_to_load_from
from which
to restore the column weights. Required if ckpt_to_load_from
is not
None
.max_norm
: If not None
, embedding values are l2-normalized to this value.trainable
: Whether or not the embedding is trainable. Default is True.DenseColumn
that converts from sparse input.
ValueError
: if dimension
not > 0.ValueError
: if exactly one of ckpt_to_load_from
and tensor_name_in_ckpt
is specified.ValueError
: if initializer
is specified and is not callable.RuntimeError
: If eager execution is enabled.