You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Added a new version of the census example to demonstrate usage in TF 2.0.
New mapper estimated_probability_density to compute either exact
probabilities (for discrete categorical variable) or approximate density over
fixed intervals (continuous variables).
New analyzers count_per_key and histogram to return counts of unique
elements or values within predefined ranges. Calling tft.histogram on
non-categorical value will assign each data point to the appropriate fixed
bucket and then count for each bucket.
Provided capability for per-key analyzers to analyze larger sets of keys that
would not fit in memory, by storing the key-value pairs in vocabulary files.
This is enabled by passing a per_key_filename to tft.scale_by_min_max_per_key and tft.scale_to_0_1_per_key.
Bug Fixes and Other Changes
Added beam counters to log analyzer and mapper usage.
Cleanup deprecated APIs used in census and sentiment examples.
Support windows style paths in analyzer_cache.
tft_beam.WriteTransformFn and tft_beam.WriteMetadata have been made
idempotent to allow retrying them in case of a failure.
tft_beam.WriteMetadata takes an optional argument write_to_unique_subdir
and returns the path to which metadata was written. If write_to_unique_subdir is True, metadata is written to a unique subdirectory
under path, otherwise it is written to path.
Support non utf-8 characters when reading vocabularies in tft.TFTransformOutput
tft.TFTransformOutput.vocabulary_by_name now returns bytes instead of str
with python 3.