Tensorflow-decision-forests

Latest version: v1.12.0

Safety actively analyzes 723158 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 4 of 6

0.2.6

Features

- Support for TensorFlow 2.9.1

0.2.5

Features

- Adds the `contrib` module for contributed, non-core functionality.
- Adds `contrib.scikit_learn_model_converter`, which facilitates converting
Scikit-Learn tree-based models into TF-DF models.
- Discard hessian splits with score lower than the parents. This change has
little effect on the model quality, but it can reduce its size.
- Add internal flag `hessian_split_score_subtract_parent` to subtract the
parent score in the computation of an hessian split score.
- Add support for hyper-parameter optimizers (also called tuner).
- Add text pretty print of trees with `tree.pretty()` or `str(tree)`.
- Add support for loading YDF models with file prefixes. Newly created models
have a random prefix attached to them. This allows combining multiple models
in Keras.
- Add support for discretized numerical features.

0.2.4

Features

- Support for TensorFlow 2.8.

0.2.3

Features

- Honest Random Forests (also work with Gradient Boosted Tree and CART).
- Can train Random Forests with example sampling without replacement.
- Add support for Focal Loss with Gradient Boosted Trees.
- Add support for MacOS.

Fixes

- Incorrect default evaluation of categorical split with uplift tasks. This
was making uplift models with missing categorical values perform worst, and
made the inference of uplift model possibly slower.
- Fix `pd_dataframe_to_tf_dataset` on Pandas dataframe not containing arrays.

0.2.2

Features

- Surface the `validation_interval_in_trees`,
`keep_non_leaf_label_distribution` and 'random_seed' hyper-parameters.
- Add the `batch_size` argument in the `pd_dataframe_to_tf_dataset` utility.
- Automatically determine the number of threads if `num_threads=None`.
- Add constructor argument `try_resume_training` to facilitate resuming
training.
- Check that the training dataset is well configured for TF-DF e.g. no repeat
operation, has a large enough batch size, etc. The check can be disabled
with `check_dataset=False`.
- When a model is created manually with the model builder, and if the dataspec
is not provided, tries to adapt the dataspec so that the model looks as if
it was trained with the global imputation strategy for missing values (i.e.
missing_value_policy: GLOBAL_IMPUTATION). This makes manually created models
more likely to be compatible with the fast inference engines.
- TF-DF models `fit` method now passes the `validation_data` to the Yggdrasil
learners. This is used for example for early stopping in the case of GBT
model.
- Add the "loss" parameter of the GBT model directly in the model constructor.
- Control the amount of training logs displayed in the notebook (if using
notebook) or in the console with the `verbose` constructor argument and
`fit` parameter of the model.

Fixes

- `num_candidate_attributes` is not ignored anymore when
`num_candidate_attributes_ratio=-1`.
- Use the median bucket split value strategy in the discretized numerical
splitters (local and distributed).
- Surface the `max_num_scanned_rows_to_accumulate_statistics` parameter to
control how many examples are scanned to determine the feature statistics
when training from a file dataset with `fit_on_dataset_path`.

0.2.1

Features

- Compatibility with TensorFlow 2.7.0.

Page 4 of 6

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.