- Depends on
pyarrow>=0.16,<0.17
.
- Added support for jackknife-based confidence intervals.
- Add EvalResult.get_metrics(), which extracts slice metrics in dictionary format from EvalResults.
- Adds TFMD
Schema
as an available argument to computations callbacks.
- Version is now available under
tfma.version.VERSION
ortfma.__version__
. - Add auto slicing utilities for significance testing.
- Fixed error when a metric and loss with the same classname are used.
- Adding two new ratios (false discovery rate and false omission rate) in Fairness Indicators.
MetricValue
s can now contain both a debug message and a value (rather than one or the other).- Fix issue with displaying ConfusionMatrixPlot in colab.
CalibrationPlot
now infersleft
andright
values from schema, when available. This makes the calibration plot useful to regression users.- Fix issue with metrics not being computed properly when mixed with specs containing micro-aggregation computations.
- Remove batched keys. Instead use the same keys for batched and unbatched extract.
- Adding support to visualize Fairness Indicators in Fairness Indicators
TensorBoard Plugin by providing remote evalution path in query parameter:
<tensorboard_url>#fairness_indicators& p.fairness_indicators.evaluation_output_path=<evaluation_path>
. - Fixed invalid metrics calculations for serving models using the classification API with binary outputs.
- Moved config writing code to extend from tfma.writer.Writer and made it a member of default_writers.
- Updated tfma.ExtractEvaluateAndWriteResults to accept Extracts as input in addition to serialize bytes and Arrow RecordBatches.
- Depends on
apache-beam[gcp]>=2.20,<3
. - Depends on
pyarrow>=0.16,<1
. - Depends on
tensorflow>=1.15,!=2.0.*,<3
. - Depends on
tensorflow-metadata>=0.22,<0.23
. - Depends on
tfx-bsl>=0.22,<0.23
.
- Remove desired_batch_size as an option. Large batch failures can be handled via serially processing the failed batch which also acts as a deterent from scaling up batch sizes further. Batch size can be handled via BEAM batch size tuning.
- Deprecating Py2 support.
- Integrate TFXIO in TFMA. Use batched input and predict extractor in V2. Results in ~40% reduction in CPU seconds over existing TFMA v2 (InputExtractor + PredictExtractorV2). Modify TFMA public API to take an optional tensor adapter config as input.
- Populate confidence_interval field in addition to bounded_value when confidence intervals is enabled.
- Only requires
avro-python3>=1.8.1,!=1.9.2.*,<2.0.0
on Python 3.5 + MacOS - Fix bug in SensitivitySpecificityBase derived metrics: guarantee well defined behaviour when the constraint lies between feasible points (see updated docstrings).
- Now publish NPM under
tensorflow_model_analysis
for UI components.
- Depends on 'tfx-bsl>=0.21.3,<0.22',
- Depends on 'tensorflow>=1.15,<3',
- Depends on 'apache-beam[gcp]>=2.17,<3',
- Rollback populating TDistributionValue metric when confidence intervals is enabled in V2.
- Added support for creating metrics specs from tf.keras.losses.
- Added evaluation comparison feature to the Fairness Indicators UI in Colab.
- Added better defaults handling for eval config so that a single model spec can be used for both candidate and baseline.
- Fixed issue with keras metrics saved with the model not being calculated unless a keras metric was added to the config.
- Depends on
pandas>=0.24,<2
. - Depends on
pyarrow>=0.15,<1
. - Depends on 'tfx-bsl>=0.21.3,<0.23',
- Depends on 'tensorflow>=1.15,!=2.0.*,<3',
- Depends on 'apache-beam[gcp]>=2.17,<2.18',
- Added support for model validation using either value threshold or diff threshold.
- Added a writer to output model validation result (ValidationResult).
- Added support for multi-model evaluation using EvalSavedModels.
- Added support for inserting model_names by default to metrics_specs.
- Added support for selecting custom model format evals in config.
- Fixed issue with model_name not being set in keras metrics.
- Populate TDistributionValue metric when confidence intervals is enabled in V2.
- Rename the writer MetricsAndPlotsWriter to MetricsPlotsAndValidationsWriter.
- Adding SciPy dependency for both Python2 and Python3
- Increased table and tooltip font in Fairness Indicators.
tfma.BinarizeOptions.class_ids
,tfma.BinarizeOptions.k_list
,tfma.BinarizeOptions.top_k_list
, andtfma.Options.disabled_outputs
are now wrapped in an additional proto message.
- Adding a TFLite predict extractor to enable obtaining inferences from TFLite models.
- Adding support to compute deterministic confidence intervals using a seed value in tfma.run_model_analysis API for testing or experimental purposes.
- Fixed calculation of
tfma.metrics.CoefficientOfDiscrimination
andtfma.metrics.RelativeCoefficientOfDiscrimination
.
- Renaming k_anonymization_count field name to min_slice_size.
- Added
tfma.metrics.MinLabelPosition
andtfma.metrics.QueryStatistics
for use with V2 metrics API. - Added
tfma.metrics.CoefficientOfDiscrimination
andtfma.metrics.RelativeCoefficientOfDiscrimination
for use with V2 metrics API. - Added support for using
tf.keras.metrics.*
metrics with V2 metrics API. - Added support for default V2 MetricSpecs and creating specs from
tf.kera.metrics.*
andtfma.metrics.*
metric classes. - Added new MetricsAndPlotsEvaluator based on V2 infrastructure. Note this evaluator also supports query-based metrics.
- Add support for micro_average, macro_average, and weighted_macro_average metrics.
- Added support for running V2 extractors and evaluators. V2 extractors will
be used whenever the default_eval_saved_model is created using a non-eval
tag (e.g.
tf.saved_model.SERVING
). The V2 evaluator will be used whenever atfma.EvalConfig
is used containingmetrics_specs
. - Added support for
tfma.metrics.SquaredPearsonCorrelation
for use with V2 metrics API. - Improved support for TPU autoscaling and handling batch_size related scaling.
- Added support for
tfma.metrics.Specificity
,tfma.metrics.FallOut
, andtfma.metrics.MissRate
for use with V2 metrics API. RenamedAUCPlot
toConfusionMatrixPlot
,MultiClassConfusionMatrixAtThresholds
toMultiClassConfusionMatrixPlot
andMultiLabelConfusionMatrixAtThresholds
toMultiLabelConfusionMatrixPlot
. - Added Jupyter support to Fairness Indicators. Currently does not support WIT integration.
- Added fairness indicators metrics
tfma.addons.fairness.metrics.FairnessIndicators
. - Updated documentation for new metrics infrastructure and newly supported models (keras, etc).
- Added support for model diff metrics. Users need to turn on "is_baseline" in the corresponding ModelSpec.
- Fixed error in
tfma-multi-class-confusion-matrix-at-thresholds
with default classNames value. - Fairness Indicators
- Compute ratio metrics with safe division.
- Remove "post_export_metrics" from metric names.
- Move threshold dropdown selector to a metric-by-metric basis, allowing different metrics to be inspected with different thresholds. Don't show thresholds for metrics that do not support them.
- Slices are now displayed in alphabetic order.
- Adding an option to "Select all" metrics in UI.
- Added auto slice key extractor based on statistics.
- Depends on 'tensorflow-metadata>=0.21,<0.22'.
- Made InputProcessor externally visible.
- Updated proto config to remove input/output data specs in favor of passing them directly to the run_eval.
- Fixed the bug that Fairness Indicator will skip metrics with NaN value.
- Updated vulcanized_tfma.js with UI changes in addons/fairness_indicators.
- Updated to use tf.io.gfile for reading config files (fixes issue with reading from GCS/HDFS in 0.15.0 and 0.15.1 releases).
- Added support for defaulting to using class IDs when classes are not present in outputs for multi-class metrics (for use in keras model_to_estimator).
- Added example count metrics (
tfma.metrics.ExampleCount
andtfma.metrics.WeightedExampleCount
) for use with V2 metrics API. - Added calibration metrics (
tfma.metrics.MeanLabel
,tfma.metrics.MeanPrediction
, andtfma.metrics.Calibration
) for use with V2 metrics API. - Added
tfma.metrics.ConfusionMatrixAtThresholds
for use with V2 metrics API. - Added
tfma.metrics.CalibrationPlot
andtfma.metrics.AUCPlot
for use with V2 metrics API. - Added multi_class / multi_label plots (
tfma.metrics.MultiClassConfusionMatrixAtThresholds
,tfma.metrics.MultiLabelConfusionMatrixAtThresholds
) for use with V2 metrics API. - Added
tfma.metrics.NDCG
metric for use with V2 metrics API. - Added
calibration
as a post export metric.
- Depends on
tensorflow>=1.15,<3.0
.- Starting from 1.15, package
tensorflow
comes with GPU support. Users won't need to choose betweentensorflow
andtensorflow-gpu
. - Caveat:
tensorflow
2.0.0 is an exception and does not have GPU support. Iftensorflow-gpu
2.0.0 is installed before installingtensorflow_model_analysis
, it will be replaced withtensorflow
2.0.0. Re-installtensorflow-gpu
2.0.0 if needed.
- Starting from 1.15, package
- Added V2 of PredictExtractor that uses TF 2.0 signature APIs and supports keras models (note: keras model evaluation not fully supported yet).
tfma.run_model_analysis
,tfma.default_extractors
,tfma.default_evaluators
, andtfma.default_writers
now allow settings to be passed as anEvalConfig
.tfma.run_model_analysis
,tfma.default_extractors
,tfma.default_evaluators
, andtfma.default_writers
now allow multiple models to be passed (note: multi-model support not fully implemented yet).- Added InputExtractor for extracting labels, features, and example weights from tf.Examples.
- Added Fairness Indicator as an addon.
- Enabled TF 2.0 support using compat.v1.
- Added support for slicing on native dicts of features in addition to FPL types.
- For multi-output and / or multi-class models, please provide output_name and / or class_id to tfma.view.render_plot.
- Replaced dependency on
tensorflow-transform
withtfx-bsl
. If running with latest master,tfx-bsl
must also be latest master. - Depends on
tfx-bsl>=0.15,<0.16
. - Slicing now supports conversion between int/floats and strings.
- Depends on
apache-beam[gcp]>=2.16,<3
. - Depends on
six==1.12
.
- tfma.EvalResult.slicing_metrics now contains nested dictionaries of output, class id and then metric names.
- Update config serialization to use JSON instead of pickling and reformat config to include input_data_specs, model_specs, output_data_specs, and metrics_specs.
- Requires pre-installed TensorFlow >=1.15,<3.
- Added documentation on architecture.
- Added an
adapt_to_remove_metrics
function totfma.exporter
which can be used to remove metrics incompatible with TFMA (e.g.py_func
or streaming metrics) before exporting the TFMA EvalSavedModel. - Added support for passing sparse int64 tensors to precision/recall@k.
- Added support for binarization of multiclass metrics that use labels of the from (N) in addition to (N, 1).
- Added support for using iterators with EvalInputReceiver.
- Improved performance of confidence interval computations by modifying the pipeline shape.
- Added QueryBasedMetricsEvaluator which supports computing query-based metrics (e.g. normalized discounted cumulative gain).
- Added support for merging metrics produced by different evaluators.
- Added support for blacklisting specified features from fetches.
- Added functionality to the FeatureExtractor to specify the features dict as a possible destination.
- Added support for label vocabularies for binary and multi-class estimators that support the new ALL_CLASSES prediction output.
- Move example parsing in aggregation into the graph for performance improvements in both standard and model_agnostic evaluation modes.
- Created separate ModelLoader type for loading the EvalSavedModel.
- Upgraded codebase for TF 2.0 compatibility.
- Make metrics-related operations thread-safe by wrapping them with locks. This eliminates race conditions that were previously possible in multi-threaded runners which could result in incorrect metric values.
- More flexible
FanoutSlices
. - Limit the number of sampling buckets to 20.
- Improved performance in Confidence Interval computation.
- Refactored poisson bootstrap code to be re-usable in other evaluators.
- Refactored k-anonymity code to be re-usable in other evaluators.
- Fixed slicer feature string value handling in Python3.
- Added support for example weight keys for multi-output models.
- Added option to set the desired batch size when calling run_model_analysis.
- Changed TFRecord compression type from UNCOMPRESSED to AUTO.
- Depends on
apache-beam[gcp]>=2.14,<3
. - Depends on
numpy>=1.16,<2
. - Depends on
protobuf>=3.7,<4
. - Depends on
scipy==1.1.0
. - Added support to change k_anonymization_count value via EvalConfig.
- Removed uses of deprecated tf.contrib packages (where possible).
tfma.default_writers
now requires theeval_saved_model
to be passed as an argument.- Requires pre-installed TensorFlow >=1.14,<2.
- Added support for squared pearson correlation (R squared) post export metric.
- Added support for mean absolute error post export metric.
- Added support for mean squared error and root mean squared error post export metric.
- Added support for not computing metrics for slices with less than a given number of examples.
- Cast / convert labels for precision / recall at K so that they work even if the label and the classes Tensors have different types, as long as the types are compatible.
- Post export metrics will now also search for prediction keys prefixed by metric_tag if it is specified.
- Added support for precision/recall @ k using canned estimators provided label vocab not used.
- Preserve unicode type of slice keys when serialising to and deserialising from disk, instead of always converting them to bytes.
- Use
__slots__
in accumulators.
- Expose Python 3 types in the code (this will break Python 2 compatibility)
- Python 3.5 is supported.
- Added support for fetching additional tensors at prediction time besides features, predictions, and labels (predict now returns FetchedTensorValues type).
- Removed internal usages of encoding.NODE_SUFFIX indirection within dicts in the eval_saved_model module (encoding.NODE_SUFFIX is still used in FeaturesPredictionLabels)
- Predictions are now returned as tensors (vs dicts) when "predictions" is the only output (this is consistent with how features and labels work).
- Depends on
apache-beam[gcp]>=2.11,<3
. - Depends on
protobuf>=3.7,<4
. - Depends on
scipy==1.1.0
. - Add support for multiple plots in a single evaluation.
- Add support for changeable confidence levels.
- Post export metrics for precision_recall_at_k were split into separate fuctions: precision_at_k and recall_at_k.
- Requires pre-installed TensorFlow >=1.13,<2.
- Python 3.5 readiness complete (all tests pass). Full Python 3.5 compatibility is expected to be available with the next version of Model Analysis (after Apache Beam 2.11 is released).
- Added support for customizing the pipeline (via extractors, evaluators, and writers). See architecture for more details.
- Added support for excluding the default metrics from the saved model graph during evaluation.
- Added a mechanism for performing evaluations via post_export_metrics without access to a Tensorflow EvalSavedModel.
- Added support for computing metrics with confidence intervals using the Poisson bootstrap technique. To use, set the num_bootstrap_samples to a number greater than 1--20 is recommended for confidence intervals.
- Fixed bugs where TFMA was incorrectly modifying elements in DoFns, which violates the Beam API.
- Fixed correctness issue stemming from TFMA incorrectly relying on evaluation ordering that TF doesn't guarantee.
- We now store feature and label Tensor information in SignatureDef inputs instead of Collections in anticipation of Collections being deprecated in TF 2.0.
- Add support for fractional labels in AUC, AUPRC and confusion matrix at
thresholds. Previously the labels were being passed directly to TensorFlow,
which would cast them to
bool
, which meant that all non-zero labels were treated as positive examples. Now we treat a fractional labell
in[0, 1]
as two examples, a positive example with weightl
and a negative example with weight1 - l
. - Depends on
numpy>=1.14.5,<2
. - Depends on
scipy==0.19.1
. - Depends on
protobuf==3.7.0rc2
. - Chicago Taxi example is moved to tfx repo (https://github.com/tensorflow/tfx/tree/master/tfx/examples/chicago_taxi)
- Moved tfma.SingleSliceSpec to tfma.slicer.SingleSliceSpec.
- We now support unsupervised models which have
model_fn
s that do not take alabels
argument. - Improved performance by using
make_callable
instead of repeatedsession.run
calls. - Improved performance by better choice of default "combine" batch size.
- We now support passing in custom extractors in the model_eval_lib API.
- Added support for models which have multiple examples per raw input (e.g.
input is a compressed example which expands to multiple examples when parsed
by the model). For such models, you must specify an
example_ref
parameter to yourEvalInputReceiver
. This 1-D integer Tensor should be batch aligned with features, predictions and labels and each element in it is an index in the raw input tensor to identify which input each feature / prediction / label came from. Seeeval_saved_model/example_trainers/fake_multi_examples_per_input_estimator.py
for an example. - Added support for metrics with string
value_op
s. - Added support for metrics whose
value_op
s return multidimensional arrays. - We now support including your serving graph in the EvalSavedModel. You can
do this by passing a
serving_input_receiver_fn
toexport_eval_savedmodel
or any of the TFMA Exporters. - We now support customizing prediction and label keys for post_export_metrics.
- Depends on
apache-beam[gcp]>=2.8,<3
. - Depends on
tensorflow-transform>=0.11,<1
. - Requires pre-installed TensorFlow >=1.11,<2.
- Factor our utility functions for adding sliceable "meta-features" to FPL.
- Added public API docs
- Add an extractor to add sliceable "meta-features" to FPL.
- Potentially improved performance by fanning out large slices.
- Add support for assets_extra in
tfma.exporter.FinalExporter
. - Add a light-weight library that includes only the export-related modules for
TFMA for use in your Trainer. See docstring in
tensorflow_model_analysis/export_only/__init__.py
for usage. - Update
EvalInputReceiver
so the TFMA collections written to the graph only contain the results of the last call if multiple calls toEvalInputReceiver
are made. - We now finalize the graph after it's loaded and post-export metrics are added, potentially improving performance.
- Fix a bug in post-export PrecisionRecallAtK where labels with only 1 dimension were not correctly handled.
- Fix an issue where we were not correctly wrapping SparseTensors for
features
andlabels
intf.identity
, which could cause TFMA to encounter TensorFlow issue #17568 if there were control dependencies on thesefeatures
orlabels
. - We now correctly preserve
dtypes
when splitting and concatenating SparseTensors internally. The failure to do so previously could result in unexpectedly large memory usage if string values were involved due to the inefficient pickling of NumPy string arrays with a large number of elements.
- Requires pre-installed TensorFlow >=1.11,<2.
- We now require that
EvalInputReceiver
,export_eval_savedmodel
,make_export_strategy
,make_final_exporter
,FinalExporter
andLatestExporter
be called with keyword arguments only. - Removed
check_metric_compatibility
fromEvalSavedModel
. - We now enforce that the
receiver_tensors
dictionary forEvalInputReceiver
contains exactly one key namedexamples
. - Post-export metrics have now been moved up one level to
tfma.post_export_metrics
. They should now be accessed viatfma.post_export_metrics.auc
instead oftfma.post_export_metrics.post_export_metrics.auc
as they were before. - Separated extraction from evaluation.
EvaluteAndWriteResults
is now calledExtractEvaluateAndWriteResults
. - Added
EvalSharedModel
type to encapsulatemodel_path
andadd_metrics_callbacks
along with a handle to a shared model instance.
- Improved performance especially when slicing across many features and/or feature values.
- Depends on
tensorflow-transform>=0.9,<1
. - Requires pre-installed TensorFlow >=1.9,<2.
- Depends on
apache-beam[gcp]>=2.6,<3
. - Updated ExampleCount to use the batch dimension as the example count. It also now tries a few fallbacks if none of the standard keys are found in the predictions dictionary: the first key in sorted order in the predictions dictionary, or failing that, the first key in sorted order in the labels dictionary, or failing that, it defaults to zero.
- Fix bug where we were mutating an element in a DoFn - this is prohibited in the Beam model and can cause subtle bugs.
- Fix bug where we were creating a separate Shared handle for each stage in Evaluate, resulting in no sharing of the model across stages.
- Requires pre-installed TensorFlow >=1.10,<2.
- Add a TFMA unit test library for unit testing your the exported model and associated metrics computations.
- Add
tfma.export.make_export_strategy
which is analogous totf.contrib.learn.make_export_strategy
. - Add
tfma.exporter.FinalExporter
andtfma.exporter.LatestExporter
which are analogous totf.estimator.FinalExporter
andtf.estimator.LastExporter
. - Add
tfma.export.build_parsing_eval_input_receiver_fn
which is analogous totf.estimator.export.build_parsing_serving_input_receiver_fn
. - Add integration testing for DNN-based estimators.
- Add new post export metrics:
- AUC (
tfma.post_export_metrics.post_export_metrics.auc
) - Precision/Recall at K
(
tfma.post_export_metrics.post_export_metrics.precision_recall_at_k
) - Confusion matrix at thresholds
(
tfma.post_export_metrics.post_export_metrics.confusion_matrix_at_thresholds
)
- AUC (
- Peak memory usage for large DataFlow jobs should be lower with a fix in when we compact batches of metrics during the combine phase of metrics computation.
- Remove batch size override in
chicago_taxi
example. - Added dependency on
protobuf>=3.6.0<4
for protocol buffers. - Updated SparseTensor code to work with SparseTensors of any dimension. Previously on SparseTensors with dimension 2 (batch_size x values) were supported in the features dictionary.
- Updated code to work with SparseTensors and dense Tensors of variable lengths across batches.
- EvalSavedModels produced by TFMA 0.6.0 will not be compatible with later
versions due to the following changes:
- EvalSavedModels are now written out with a custom "eval_saved_model" tag, as opposed to the "serving" tag before.
- EvalSavedModels now include version metadata about the TFMA version that they were exported with.
- Metrics and plot outputs now are converted into proto and serialized. Metrics and plots produced by TFMA 0.6.0 will not be compatible with later versions.
- Requires pre-installed TensorFlow >=1.9,<2.
- TFMA now uses the TensorFlow Estimator functionality for exporting models of different modes behind the scenes. There are no user-facing changes API-wise, but EvalSavedModels produced by earlier versions of TFMA will not be compatible with this version of TFMA.
- tf.contrib.learn Estimators are no longer supported by TFMA. Only tf.estimator Estimators are supported.
- Metrics and plot outputs now include version metadata about the TFMA version that they were exported with. Metrics and plots produced by earlier versions of TFMA will not be compatible with this version of TFMA.
- Initial release of TensorFlow Model Analysis.