+ ```bash
+ tfx template copy --model=template --pipeline_name=pipeline-name \
+ --destination_path=destination-path
+ ```
Replace the following:
- * template: The name of the template you want to copy.
- * pipeline-name: The name of the pipeline to create.
- * destination-path: The path to copy the template into.
+ * `template`: The name of the template you want to copy.
+ * `pipeline-name`: The name of the pipeline to create.
+ * `destination-path`: The path to copy the template into.
- Learn more about the [`tfx template copy` command](cli#copy).
+ Learn more about the [`tfx template copy` command](cli.md#copy).
1. A copy of the pipeline template has been created at the path you specified.
@@ -99,13 +99,13 @@ This section provides an overview of the scaffolding created by a template.
1. Run the following commands in your pipeline directory:
-
+ ```bash
tfx run create --pipeline_name pipeline_name
-
+ ```
The command creates a pipeline run using `LocalDagRunner`, which adds the
following directories to your pipeline:
@@ -157,8 +157,8 @@ template.
implement a pipeline for tabular data using the TFX standard components. If
you are moving an existing ML workflow into a pipeline, you may need to
revise your code to make full use of
- [TFX standard components](index#tfx_standard_components). You may also need
- to create [custom components](understanding_custom_components) that
+ [TFX standard components](index.md#tfx_standard_components). You may also need
+ to create [custom components](understanding_custom_components.md) that
implement features which are unique to your workflow or that are not yet
supported by TFX standard components.
@@ -194,17 +194,17 @@ without using a template.
functionality to help you implement a complete ML workflow. If you are
moving an existing ML workflow into a pipeline, you may need to revise your
code to make full use of TFX standard components. You may also need to
- create [custom components](understanding_custom_components) that implement
+ create [custom components](understanding_custom_components.md) that implement
features such as data augmentation.
* Learn more about
- [standard TFX components](index#tfx_standard_components).
- * Learn more about [custom components](understanding_custom_components).
+ [standard TFX components](index.md#tfx_standard_components).
+ * Learn more about [custom components](understanding_custom_components.md).
1. Create a script file to define your pipeline using the following example.
This guide refers to this file as `my_pipeline.py`.
-
+ ```python
import os
from typing import Optional, Text, List
from absl import logging
@@ -248,7 +248,7 @@ without using a template.
if __name__ == '__main__':
logging.set_verbosity(logging.INFO)
run_pipeline()
-
+ ```
In the coming steps, you define your pipeline in `create_pipeline` and run
your pipeline locally using the local runner.
@@ -277,7 +277,7 @@ without using a template.
pipeline uses the `ExampleGen` standard component to load a CSV from a
directory at `./data`.
-
+ ```python
from tfx.components import CsvExampleGen
DATA_PATH = os.path.join('.', 'data')
@@ -315,7 +315,7 @@ without using a template.
)
tfx.orchestration.LocalDagRunner().run(my_pipeline)
-
+ ```
`CsvExampleGen` creates serialized example records using the data in the CSV
at the specified data path. By setting the `CsvExampleGen` component's
@@ -326,13 +326,13 @@ without using a template.
1. Use the following command to run your `my_pipeline.py` script.
-
+ ```bash
python my_pipeline.py
-
+ ```
The result should be something like the following:
-
+ ```
INFO:absl:Component CsvExampleGen depends on [].
INFO:absl:Component CsvExampleGen is scheduled.
INFO:absl:Component CsvExampleGen is running.
@@ -347,6 +347,6 @@ without using a template.
INFO:absl:Running publisher for CsvExampleGen
INFO:absl:MetadataStore with DB connection initialized
INFO:absl:Component CsvExampleGen is finished.
-
+ ```
1. Continue to iteratively add components to your pipeline.
diff --git a/docs/guide/build_tfx_pipeline.md b/docs/guide/build_tfx_pipeline.md
index 5cfbe0f85b..f03a5f4648 100644
--- a/docs/guide/build_tfx_pipeline.md
+++ b/docs/guide/build_tfx_pipeline.md
@@ -1,11 +1,11 @@
# Building TFX pipelines
Note: For a conceptual view of TFX Pipelines, see
-[Understanding TFX Pipelines](understanding_tfx_pipelines).
+[Understanding TFX Pipelines](understanding_tfx_pipelines.md).
Note: Want to build your first pipeline before you dive into the details? Get
started
-[building a pipeline using a template](https://www.tensorflow.org/tfx/guide/build_local_pipeline#build_a_pipeline_using_a_template).
+[building a pipeline using a template](build_local_pipeline.md#build-a-pipeline-using-a-template).
## Using the `Pipeline` class
@@ -13,37 +13,37 @@ TFX pipelines are defined using the
[`Pipeline` class](https://github.com/tensorflow/tfx/blob/master/tfx/orchestration/pipeline.py){: .external }.
The following example demonstrates how to use the `Pipeline` class.
-
+```
Replace the following:
-* pipeline-name: The name of this pipeline. The pipeline name must
+* `pipeline-name`: The name of this pipeline. The pipeline name must
be unique.
TFX uses the pipeline name when querying ML Metadata for component input
artifacts. Reusing a pipeline name may result in unexpected behaviors.
-* pipeline-root: The root path of this pipeline's outputs. The root
+* `pipeline-root`: The root path of this pipeline's outputs. The root
path must be the full path to a directory that your orchestrator has read
and write access to. At runtime, TFX uses the pipeline root to generate
output paths for component artifacts. This directory can be local, or on a
supported distributed file system, such as Google Cloud Storage or HDFS.
-* components: A list of component instances that make up this
+* `components`: A list of component instances that make up this
pipeline's workflow.
-* enable-cache: (Optional.) A boolean value that indicates if this
+* `enable-cache`: (Optional.) A boolean value that indicates if this
pipeline uses caching to speed up pipeline execution.
-* metadata-connection-config: (Optional.) A connection
+* `metadata-connection-config`: (Optional.) A connection
configuration for ML Metadata.
## Defining the component execution graph
diff --git a/docs/guide/bulkinferrer.md b/docs/guide/bulkinferrer.md
index e96735d014..9b5e364d55 100644
--- a/docs/guide/bulkinferrer.md
+++ b/docs/guide/bulkinferrer.md
@@ -2,7 +2,7 @@
The BulkInferrer TFX component performs batch inference on unlabeled data. The
generated
-InferenceResult([tensorflow_serving.apis.prediction_log_pb2.PredictionLog](https://github.com/tensorflow/serving/blob/master/tensorflow_serving/apis/prediction_log.proto))
+InferenceResult([`tensorflow_serving.apis.prediction_log_pb2.PredictionLog`](https://github.com/tensorflow/serving/blob/master/tensorflow_serving/apis/prediction_log.proto))
contains the original features and the prediction results.
BulkInferrer consumes:
@@ -11,7 +11,7 @@ BulkInferrer consumes:
[SavedModel](https://www.tensorflow.org/guide/saved_model.md) format.
* Unlabelled tf.Examples that contain features.
* (Optional) Validation result from
- [Evaluator](https://www.tensorflow.org/tfx/guide/evaluator.md) component.
+ [Evaluator](evaluator.md) component.
BulkInferrer emits:
@@ -21,9 +21,9 @@ BulkInferrer emits:
A BulkInferrer TFX component is used to perform batch inference on unlabeled
tf.Examples. It is typically deployed after an
-[Evaluator](https://www.tensorflow.org/tfx/guide/evaluator.md) component to
+[Evaluator](evaluator.md) component to
perform inference with a validated model, or after a
-[Trainer](https://www.tensorflow.org/tfx/guide/trainer.md) component to directly
+[Trainer](trainer.md) component to directly
perform inference on exported model.
It currently performs in-memory model inference and remote inference.
@@ -42,4 +42,4 @@ bulk_inferrer = BulkInferrer(
```
More details are available in the
-[BulkInferrer API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/BulkInferrer).
+[BulkInferrer API reference][tfx.v1.components.BulkInferrer].
diff --git a/docs/guide/cli.md b/docs/guide/cli.md
index 46fa26a138..855f5d2bdd 100644
--- a/docs/guide/cli.md
+++ b/docs/guide/cli.md
@@ -18,19 +18,19 @@ interface might change as new versions are released.
The TFX CLI is installed as a part of the TFX package. All CLI commands follow
the structure below:
-
-tfx command-groupcommandflags
-
+```bash
+tfx
+```
-The following command-group options are currently supported:
+The following command-group options are currently supported:
-* [tfx pipeline](#tfx-pipeline) - Create and manage TFX pipelines.
-* [tfx run](#tfx-run) - Create and manage runs of TFX pipelines on various
+* [`tfx pipeline`](#tfx-pipeline) - Create and manage TFX pipelines.
+* [`tfx run`](#tfx-run) - Create and manage runs of TFX pipelines on various
orchestration platforms.
-* [tfx template](#tfx-template-experimental) - Experimental commands for
+* [`tfx template`](#tfx-template-experimental) - Experimental commands for
listing and copying TFX pipeline templates.
-Each command group provides a set of commands. Follow the
+Each command group provides a set of commands. Follow the
instructions in the [pipeline commands](#tfx-pipeline),
[run commands](#tfx-run), and [template commands](#tfx-template-experimental)
sections to learn more about using these commands.
@@ -42,15 +42,15 @@ Flags let you pass arguments into CLI commands. Words in flags are separated
with either a hyphen (`-`) or an underscore (`_`). For example, the pipeline
name flag can be specified as either `--pipeline-name` or `--pipeline_name`.
This document specifies flags with underscores for brevity. Learn more about
-[flags used in the TFX CLI](#understanding-tfx-cli-flags).
+[flags used in the TFX CLI](#understanding-tfx-cli-flags).
## tfx pipeline
The structure for commands in the `tfx pipeline` command group is as follows:
-
+```bash
+tfx pipeline command required-flags [optional-flags]
+```
Use the following sections to learn more about the commands in the `tfx
pipeline` command group.
@@ -61,11 +61,11 @@ Creates a new pipeline in the given orchestrator.
Usage:
-
+```
To autodetect engine from user environment, simply avoid using the engine flag
like the example below. For more details, check the flags section.
-
+```
### list
@@ -466,10 +466,10 @@ Lists all the pipelines in the given orchestrator.
Usage:
-
+```bash
tfx pipeline list [--endpoint=endpoint --engine=engine \
--iap_client_id=iap-client-id --namespace=namespace]
-
+```
--endpoint=endpoint
@@ -533,34 +533,34 @@ tfx pipeline list [--endpoint=endpoint --engine=engine \
-#### Examples:
+#### Examples
Kubeflow:
-
+```bash
tfx pipeline list --engine=kubeflow --iap_client_id=iap-client-id \
--namespace=namespace --endpoint=endpoint
-
+```
Local:
-
+```bash
tfx pipeline list --engine=local
-
+```
Vertex:
-
+```bash
tfx pipeline list --engine=vertex
-
+```
## tfx run
The structure for commands in the `tfx run` command group is as follows:
-
+```bash
tfx run commandrequired-flags [optional-flags]
-
+```
Use the following sections to learn more about the commands in the `tfx run`
command group.
@@ -572,10 +572,10 @@ most recent pipeline version of the pipeline in the cluster is used.
Usage:
-
+```
Use the following sections to learn more about the commands in the `tfx
template` command group. Template is an experimental feature and subject to
@@ -1033,9 +1033,9 @@ List available TFX pipeline templates.
Usage:
-
+```bash
tfx template list
-
+```
### copy
@@ -1043,10 +1043,10 @@ Copy a template to the destination directory.
Usage:
-
diff --git a/docs/guide/container_component.md b/docs/guide/container_component.md
index 4deb61e786..67449cc7b9 100644
--- a/docs/guide/container_component.md
+++ b/docs/guide/container_component.md
@@ -5,7 +5,7 @@ any language into your pipeline, so long as you can execute that code in a
Docker container.
If you are new to TFX pipelines,
-[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines).
+[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines.md).
## Creating a Container-based Component
diff --git a/docs/guide/custom_component.md b/docs/guide/custom_component.md
index f9c12ca41f..9527f3bbe2 100644
--- a/docs/guide/custom_component.md
+++ b/docs/guide/custom_component.md
@@ -6,7 +6,7 @@ specification, executor, and component interface classes. This approach lets you
reuse and extend a standard component to fit your needs.
If you are new to TFX pipelines,
-[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines).
+[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines.md).
## Custom executor or custom component
diff --git a/docs/guide/custom_function_component.md b/docs/guide/custom_function_component.md
index 432ad28215..8aca8be9aa 100644
--- a/docs/guide/custom_function_component.md
+++ b/docs/guide/custom_function_component.md
@@ -64,7 +64,7 @@ def MyDataProcessor(
```
If you are new to TFX pipelines,
-[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines).
+[learn more about the core concepts of TFX pipelines](understanding_tfx_pipelines.md).
## Inputs, outputs, and parameters
diff --git a/docs/guide/evaluator.md b/docs/guide/evaluator.md
index ed99871521..a1a72ab15e 100644
--- a/docs/guide/evaluator.md
+++ b/docs/guide/evaluator.md
@@ -15,7 +15,7 @@ the [Pusher](pusher.md) that it is ok to push the model to production.
* Consumes:
* An eval split from
- [Examples](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/types/standard_artifacts/Examples)
+ [Examples][tfx.v1.types.standard_artifacts.Examples]
* A trained model from [Trainer](trainer.md)
* A previously blessed model (if validation to be performed)
* Emits:
@@ -142,4 +142,4 @@ if not validation_result.validation_ok:
```
More details are available in the
-[Evaluator API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/Evaluator).
+[Evaluator API reference][tfx.v1.components.Evaluator].
diff --git a/docs/guide/examplegen.md b/docs/guide/examplegen.md
index 9f4712fdb8..aff3284de2 100644
--- a/docs/guide/examplegen.md
+++ b/docs/guide/examplegen.md
@@ -34,7 +34,7 @@ components for these data sources and formats:
* [Parquet](https://github.com/tensorflow/tfx/blob/master/tfx/components/example_gen/custom_executors/parquet_executor.py)
See the usage examples in the source code and
-[this discussion](/tfx/guide/examplegen#custom_examplegen) for more information on
+[this discussion](examplegen.md#custom_examplegen) for more information on
how to use and develop custom executors.
Note: In most case it's better to inherit from `base_example_gen_executor`
@@ -42,7 +42,7 @@ instead of `base_executor`. So following the Avro or Parquet example in the
Executor source code may be advisable.
In addition, these data sources and formats are available as
-[custom component](/tfx/guide/understanding_custom_components) examples:
+[custom component](understanding_custom_components.md) examples:
* [Presto](https://github.com/tensorflow/tfx/tree/master/tfx/examples/custom_components/presto_example_gen)
@@ -629,7 +629,7 @@ evaluator = Evaluator(
```
More details are available in the
-[CsvExampleGen API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/CsvExampleGen),
-[FileBasedExampleGen API implementation](https://github.com/tensorflow/tfx/blob/master/tfx/components/example_gen/component.py)
+[CsvExampleGen API reference][tfx.v1.components.CsvExampleGen],
+[FileBasedExampleGen API implementation][tfx.v1.components.example_gen.component],
and
-[ImportExampleGen API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/ImportExampleGen).
+[ImportExampleGen API reference][tfx.v1.components/ImportExampleGen].
diff --git a/docs/guide/exampleval.md b/docs/guide/exampleval.md
index 3f9c6ef949..e41823373e 100644
--- a/docs/guide/exampleval.md
+++ b/docs/guide/exampleval.md
@@ -38,4 +38,4 @@ validate_stats = ExampleValidator(
```
More details are available in the
-[ExampleValidator API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/ExampleValidator).
+[ExampleValidator API reference][tfx.v1.components.ExampleValidator].
diff --git a/docs/guide/fairness_indicators.md b/docs/guide/fairness_indicators.md
index 785faab5f9..88192873ae 100644
--- a/docs/guide/fairness_indicators.md
+++ b/docs/guide/fairness_indicators.md
@@ -51,7 +51,7 @@ model, please see the “Model-Agnostic TFMA” section below.
After your Estimator is trained, you will need to export a saved model for
evaluation purposes. To learn more, see the
-[TFMA guide](/tfx/model_analysis/get_started).
+[TFMA guide](https://www.tensorflow.org/tfx/model_analysis/get_started).
### Configuring Slices
diff --git a/docs/guide/index.md b/docs/guide/index.md
index 4af4795144..dd1001ca38 100644
--- a/docs/guide/index.md
+++ b/docs/guide/index.md
@@ -62,19 +62,19 @@ environment. TFX provides the following:
ML workflow on several platforms, such as: Apache Airflow, Apache Beam, and
Kubeflow Pipelines.
- [Learn more about TFX pipelines](https://www.tensorflow.org/tfx/guide/understanding_tfx_pipelines).
+ [Learn more about TFX pipelines](understanding_tfx_pipelines.md).
* A set of standard components that you can use as a part of a pipeline, or as
a part of your ML training script. TFX standard components provide proven
functionality to help you get started building an ML process easily.
- [Learn more about TFX standard components](#tfx_standard_components).
+ [Learn more about TFX standard components](#tfx-standard-components).
* Libraries which provide the base functionality for many of the standard
components. You can use the TFX libraries to add this functionality to your
own custom components, or use them separately.
- [Learn more about the TFX libraries](#tfx_libraries).
+ [Learn more about the TFX libraries](#tfx-libraries).
TFX is a Google-production-scale machine learning toolkit based on TensorFlow.
It provides a configuration framework and shared libraries to integrate common
@@ -412,7 +412,7 @@ A typical TFX pipeline will include a [Transform](transform.md) component, which
will perform feature engineering by leveraging the capabilities of the
[TensorFlow Transform (TFT)](tft.md) library. A Transform component consumes the
schema created by a SchemaGen component, and applies
-[data transformations](https://www.tensorflow.org/tfx/tutorials/transform/simple)
+[data transformations](../tutorials/transform/simple)
to create, combine, and transform the features that will be used to train your
model. Cleanup of missing values and conversion of types should also be done in
the Transform component if there is ever a possibility that these will also be
@@ -568,7 +568,7 @@ on using TensorFlow JS.
## Creating a TFX Pipeline With Airflow
Check
-[airflow workshop](https://www.tensorflow.org/tfx/tutorials/tfx/airflow_workshop/)
+[airflow workshop](../tutorials/tfx/airflow_workshop/)
for details
## Creating a TFX Pipeline With Kubeflow
@@ -582,7 +582,7 @@ Kubeflow deployment guideline that guide through the options for
### Configure and run TFX pipeline
Please follow the
-[TFX on Cloud AI Platform Pipeline tutorial](https://www.tensorflow.org/tfx/tutorials/tfx/cloud-ai-platform-pipelines)
+[TFX on Cloud AI Platform Pipeline tutorial](../tutorials/tfx/cloud-ai-platform-pipelines/)
to run the TFX example pipeline on Kubeflow. TFX components have been
containerized to compose the Kubeflow pipeline and the sample illustrates the
ability to configure the pipeline to read large public dataset and execute
diff --git a/docs/guide/infra_validator.md b/docs/guide/infra_validator.md
index 021026997c..1daeea2856 100644
--- a/docs/guide/infra_validator.md
+++ b/docs/guide/infra_validator.md
@@ -54,7 +54,7 @@ modes:
Usually InfraValidator is defined next to an Evaluator component, and its output
is fed to a Pusher. If InfraValidator fails, the model will not be pushed.
-```python {highlight="lines:8-11 context:infra_blessing,1"}
+```python hl_lines="8-11"
evaluator = Evaluator(
model=trainer.outputs['model'],
examples=example_gen.outputs['examples'],
@@ -108,7 +108,7 @@ block of the `ServingSpec`. For example to use TensorFlow Serving binary running
on the Kubernetes cluster, `tensorflow_serving` and `kubernetes` field should be
set.
-```python {highlight="lines:4:9-4:26,7:9-7:18"}
+```python hl_lines="4 7"
infra_validator=InfraValidator(
model=trainer.outputs['model'],
serving_spec=tfx.proto.ServingSpec(
@@ -127,7 +127,7 @@ To further configure `ServingSpec`, please check out the
Optional configuration to adjust the infra validation criteria or workflow.
-```python {highlight="lines:4-10"}
+```python hl_lines="4-10"
infra_validator=InfraValidator(
model=trainer.outputs['model'],
serving_spec=tfx.proto.ServingSpec(...),
@@ -151,7 +151,7 @@ infra validation in `LOAD_AND_QUERY` mode. In order to use `LOAD_AND_QUERY`
mode, it is required to specify both `request_spec` execution properties as well
as `examples` input channel in the component definition.
-```python {highlight="lines:7:9-7:62 lines:10-16"}
+```python hl_lines="8 11-17"
infra_validator = InfraValidator(
model=trainer.outputs['model'],
# This is the source for the data that will be used to build a request.
@@ -198,7 +198,7 @@ and can also be pushed by the [Pusher](pusher.md), just like `Model` artifact.
Current InfraValidator is not complete yet, and has some limitations.
-- Only TensorFlow [SavedModel](/guide/saved_model) model format can be
+- Only TensorFlow [SavedModel](https://www.tensorflow.org/guide/saved_model) model format can be
validated.
- When running TFX on Kubernetes, the pipeline should be executed by
`KubeflowDagRunner` inside Kubeflow Pipelines. The model server will be
@@ -206,13 +206,13 @@ Current InfraValidator is not complete yet, and has some limitations.
using.
- InfraValidator is primarily focused on deployments to
[TensorFlow Serving](serving.md), and while still useful it is less accurate
- for deployments to [TensorFlow Lite](/lite) and [TensorFlow.js](/js), or
+ for deployments to [TensorFlow Lite](https://www.tensorflow.org/lite) and [TensorFlow.js](https://www.tensorflow.org/js), or
other inference frameworks.
- There's a limited support on `LOAD_AND_QUERY` mode for the
[Predict](/versions/r1.15/api_docs/python/tf/saved_model/predict_signature_def)
method signature (which is the only exportable method in TensorFlow 2).
InfraValidator requires the Predict signature to consume a serialized
- [`tf.Example`](/tutorials/load_data/tfrecord#tfexample) as the only input.
+ [`tf.Example`](https://www.tensorflow.org/tutorials/load_data/tfrecord#tfexample) as the only input.
```python
@tf.function
diff --git a/docs/guide/keras.md b/docs/guide/keras.md
index 275a3bd61c..dd1454db9a 100644
--- a/docs/guide/keras.md
+++ b/docs/guide/keras.md
@@ -106,7 +106,7 @@ Here are several examples with native Keras:
end-to-end example with advanced Transform usage.
We also have a per-component
-[Keras Colab](https://www.tensorflow.org/tfx/tutorials/tfx/components_keras).
+[Keras Colab](../../tutorials/tfx/components_keras).
### TFX Components
diff --git a/docs/guide/kubeflow.md b/docs/guide/kubeflow.md
index ad94a26c64..e29b531851 100644
--- a/docs/guide/kubeflow.md
+++ b/docs/guide/kubeflow.md
@@ -15,5 +15,5 @@ Pipelines SDK allows for creation and sharing of components and composition and
of pipelines programmatically.
See the
-[TFX example on Kubeflow Pipelines](https://www.tensorflow.org/tfx/tutorials/tfx/cloud-ai-platform-pipelines)
+[TFX example on Kubeflow Pipelines](../../tutorials/tfx/cloud-ai-platform-pipelines)
for details on running TFX at scale on Google cloud.
diff --git a/docs/guide/local_orchestrator.md b/docs/guide/local_orchestrator.md
index 74bd5c6fb3..049a2e2421 100644
--- a/docs/guide/local_orchestrator.md
+++ b/docs/guide/local_orchestrator.md
@@ -5,8 +5,8 @@
Local orchestrator is a simple orchestrator that is included in the TFX Python
package. It runs pipelines in the local environment in a single process. It
provides fast iterations for development and debugging, but it is not suitable for
-large production workloads. Please use [Vertex Pipelines](/tfx/guide/vertex) or
-[Kubeflow Pipelines](/tfx/guide/kubeflow) for production use cases.
+large production workloads. Please use [Vertex Pipelines](vertex.md) or
+[Kubeflow Pipelines](kubeflow.md) for production use cases.
-Try the [TFX tutorials](/tfx/tutorials/tfx/penguin_simple) running in Colab to
+Try the [TFX tutorials](../../tutorials/tfx/penguin_simple) running in Colab to
learn how to use the local orchestrator.
diff --git a/docs/guide/mlmd.md b/docs/guide/mlmd.md
index a283e1f7a3..b2cdb58973 100644
--- a/docs/guide/mlmd.md
+++ b/docs/guide/mlmd.md
@@ -191,7 +191,7 @@ following list provides a non-exhaustive overview of some of the major benefits.
within a range; find previous executions in a context with the same inputs.
See the
-[MLMD tutorial](https://www.tensorflow.org/tfx/tutorials/mlmd/mlmd_tutorial) for
+[MLMD tutorial](../../tutorials/mlmd/mlmd_tutorial) for
an example that shows you how to use the MLMD API and the metadata store to
retrieve lineage information.
@@ -439,7 +439,7 @@ to learn how to use MLMD declarative nodes filtering capabilities on properties
and 1-hop neighborhood nodes.
Also check out the
-[MLMD tutorial](https://www.tensorflow.org/tfx/tutorials/mlmd/mlmd_tutorial) to
+[MLMD tutorial](../../tutorials/mlmd/mlmd_tutorial) to
learn how to use MLMD to trace the lineage of your pipeline components.
MLMD provides utilities to handle schema and data migrations across releases.
diff --git a/docs/guide/non_tf.md b/docs/guide/non_tf.md
index 1727bb4c7f..0bfde25fc3 100644
--- a/docs/guide/non_tf.md
+++ b/docs/guide/non_tf.md
@@ -32,7 +32,7 @@ using the standard TFX components with other frameworks include:
instead of raw features, and users can run transform as a preprocessing
step before calling the model prediction when serving.
* **Trainer** supports
- [GenericTraining](https://www.tensorflow.org/tfx/guide/trainer#generic_trainer)
+ [GenericTraining](trainer.md#generic-trainer)
so users can train their models using any ML framework.
* **Evaluator** by default only supports `saved_model`, but users can provide
a UDF that generates predictions for model evaluation.
@@ -49,7 +49,7 @@ high-performance machine learning research.
is a neural network library and ecosystem for JAX, designed for flexibility.
With [jax2tf](https://github.com/google/jax/tree/main/jax/experimental/jax2tf),
-we are able to convert trained JAX/Flax models into `saved_model` format,
+we are able to convert trained JAX/Flax models into `saved_model` format,
which can be used seamlessly in TFX with generic training and model evaluation.
For details, check this [example](https://github.com/tensorflow/tfx/blob/master/tfx/examples/penguin/penguin_utils_flax_experimental.py).
diff --git a/docs/guide/pusher.md b/docs/guide/pusher.md
index 1b3b386f7c..8b68f73727 100644
--- a/docs/guide/pusher.md
+++ b/docs/guide/pusher.md
@@ -1,16 +1,16 @@
# The Pusher TFX Pipeline Component
The Pusher component is used to push a validated model to a
-[deployment target](index.md#deployment_targets) during model training or
+[deployment target](index.md#deployment-targets) during model training or
re-training. Before the deployment, Pusher relies on one or more blessings from
other validation components to decide whether to push the model or not.
-- [Evaluator](evaluator) blesses the model if the new trained model is "good
+- [Evaluator](evaluator.md) blesses the model if the new trained model is "good
enough" to be pushed to production.
-- (Optional but recommended) [InfraValidator](infra_validator) blesses the
+- (Optional but recommended) [InfraValidator](infra_validator.md) blesses the
model if the model is mechanically servable in a production environment.
-A Pusher component consumes a trained model in [SavedModel](/guide/saved_model)
+A Pusher component consumes a trained model in [SavedModel](https://www.tensorflow.org/guide/saved_model)
format, and produces the same SavedModel, along with versioning metadata.
## Using the Pusher Component
@@ -36,7 +36,7 @@ pusher = Pusher(
(From version 0.30.0)
InfraValidator can also produce `InfraBlessing` artifact containing a
-[model with warmup](infra_validator#producing_a_savedmodel_with_warmup), and
+[model with warmup](infra_validator.md#producing-a-savedmodel-with-warmup), and
Pusher can push it just like a `Model` artifact.
```python
@@ -55,4 +55,4 @@ pusher = Pusher(
```
More details are available in the
-[Pusher API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/Pusher).
+[Pusher API reference][tfx.v1.components.Pusher].
diff --git a/docs/guide/schemagen.md b/docs/guide/schemagen.md
index d1fd36230d..2bbd50b0fe 100644
--- a/docs/guide/schemagen.md
+++ b/docs/guide/schemagen.md
@@ -58,7 +58,7 @@ The modified schema can be brought back into the pipeline using ImportSchemaGen
component. The SchemaGen component for the initial schema generation can be
removed and all downstream components can use the output of ImportSchemaGen. It
is also recommended to add
-[ExampleValidator](https://www.tensorflow.org/tfx/guide/exampleval) using the
+[ExampleValidator](exampleval.md) using the
imported schema to examine the training data continuously.
## SchemaGen and TensorFlow Data Validation
@@ -78,7 +78,7 @@ schema_gen = tfx.components.SchemaGen(
```
More details are available in the
-[SchemaGen API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/SchemaGen).
+[SchemaGen API reference][tfx.v1.components.SchemaGen].
### For the reviewed schema import
@@ -93,4 +93,4 @@ schema_gen = tfx.components.ImportSchemaGen(
The `schema_file` should be a full path to the text protobuf file.
More details are available in the
-[ImportSchemaGen API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/ImportSchemaGen).
+[ImportSchemaGen API reference][tfx.v1.components.ImportSchemaGen].
diff --git a/docs/guide/solutions.md b/docs/guide/solutions.md
index 0f8f9e9da1..f14b6fb47f 100644
--- a/docs/guide/solutions.md
+++ b/docs/guide/solutions.md
@@ -18,8 +18,7 @@ understand what items your customers consider to be similar, which enables you
to offer real-time "similar item" suggestions in your application. This solution
shows you how to identify similar songs in a dataset, and then use this
information to make song recommendations.
-Read
-more
+[Read more](https://cloud.google.com/solutions/real-time-item-matching)
## Data preprocessing for machine learning: options and recommendations
@@ -31,10 +30,8 @@ article focuses on using TensorFlow and the open source TensorFlow Transform
prediction. This part highlights the challenges of preprocessing data for
machine learning, and illustrates the options and scenarios for performing data
transformation on Google Cloud effectively.
-Part
-1
-Part
-2
+[Part 1](https://cloud.google.com/solutions/machine-learning/data-preprocessing-for-ml-with-tf-transform-pt1)
+[Part 2](https://cloud.google.com/solutions/machine-learning/data-preprocessing-for-ml-with-tf-transform-pt2)
## Architecture for MLOps using TFX, Kubeflow Pipelines, and Cloud Build
@@ -42,8 +39,7 @@ This document describes the overall architecture of a machine learning (ML)
system using TensorFlow Extended (TFX) libraries. It also discusses how to set
up a continuous integration (CI), continuous delivery (CD), and continuous
training (CT) for the ML system using Cloud Build and Kubeflow Pipelines.
-Read
-more
+[Read more](https://cloud.google.com/solutions/machine-learning/architecture-for-mlops-using-tfx-kubeflow-pipelines-and-cloud-build)
## MLOps: Continuous delivery and automation pipelines in machine learning
@@ -52,8 +48,7 @@ integration (CI), continuous delivery (CD), and continuous training (CT) for
machine learning (ML) systems. Data science and ML are becoming core
capabilities for solving complex real-world problems, transforming industries,
and delivering value in all domains.
-Read
-more
+[Read more](https://cloud.google.com/solutions/machine-learning/mlops-continuous-delivery-and-automation-pipelines-in-machine-learning)
## Setting up an MLOps environment on Google Cloud
@@ -64,8 +59,7 @@ environment described here. Virtually all industries are adopting machine
learning (ML) at a rapidly accelerating pace. A key challenge for getting value
from ML is to create ways to deploy and operate ML systems effectively. This
guide is intended for machine learning (ML) and DevOps engineers.
-Read
-more
+[Read more](https://cloud.google.com/solutions/machine-learning/setting-up-an-mlops-environment)
## Key requirements for an MLOps foundation
@@ -78,8 +72,7 @@ McKinsey Global Institute.
But it’s not easy right now. Machine learning (ML) systems have a special
capacity for creating technical debt if not managed well.
-Read
-more
+[Read more](https://cloud.google.com/blog/products/ai-machine-learning/key-requirements-for-an-mlops-foundation)
## How to create and deploy a model card in the cloud with Scikit-Learn
@@ -88,8 +81,7 @@ With their vast potential, ML models also raise questions about their usage,
construction, and limitations. Documenting the answers to these questions helps
to bring clarity and shared understanding. To help advance these goals, Google
has introduced model cards.
-Read
-more
+[Read more](https://cloud.google.com/blog/products/ai-machine-learning/create-a-model-card-with-scikit-learn)
## Analyzing and validating data at scale for machine learning with TensorFlow Data Validation
@@ -99,5 +91,4 @@ scientists and machine learning (ML) engineers can use TFDV in a production ML
system to validate data that's used in a continuous training (CT) pipeline, and
to detect skews and outliers in data received for prediction serving. It
includes **hands-on labs**.
-Read
-more
+[Read more](https://cloud.google.com/solutions/machine-learning/analyzing-and-validating-data-at-scale-for-ml-using-tfx)
diff --git a/docs/guide/statsgen.md b/docs/guide/statsgen.md
index 7d734fa4f6..04ad7a4fa5 100644
--- a/docs/guide/statsgen.md
+++ b/docs/guide/statsgen.md
@@ -64,8 +64,8 @@ Where `` represents a unique ID for this version of the schema in
MLMD. This schema proto can then be modified to communicate information about
the dataset which cannot be reliably inferred, which will make the output of
`StatisticsGen` more useful and the validation performed in the
-[`ExampleValidator`](https://www.tensorflow.org/tfx/guide/exampleval) component
+[`ExampleValidator`](exampleval.md) component
more stringent.
More details are available in the
-[StatisticsGen API reference](https://www.tensorflow.org/tfx/api_docs/python/tfx/v1/components/StatisticsGen).
+[StatisticsGen API reference][tfx.v1.components.StatisticsGen].
diff --git a/docs/guide/tfdv.md b/docs/guide/tfdv.md
index 938ef2e261..b496170d86 100644
--- a/docs/guide/tfdv.md
+++ b/docs/guide/tfdv.md
@@ -24,9 +24,9 @@ TFX tools can both help find data bugs, and help with feature engineering.
## TensorFlow Data Validation
* [Overview](#overview)
-* [Schema Based Example Validation](#schema_based_example_validation)
+* [Schema Based Example Validation](#schema_based-example-validation)
* [Training-Serving Skew Detection](#skewdetect)
-* [Drift Detection](#drift_detection)
+* [Drift Detection](#drift-detection)
### Overview
@@ -42,9 +42,9 @@ be configured to detect different classes of anomalies in the data. It can
We document each of these functionalities independently:
-* [Schema Based Example Validation](#schema_based_example_validation)
+* [Schema Based Example Validation](#schema_based-example-validation)
* [Training-Serving Skew Detection](#skewdetect)
-* [Drift Detection](#drift_detection)
+* [Drift Detection](#drift-detection)
### Schema Based Example Validation
diff --git a/docs/guide/tfma.md b/docs/guide/tfma.md
index be7380ff7a..6facaa1e06 100644
--- a/docs/guide/tfma.md
+++ b/docs/guide/tfma.md
@@ -15,25 +15,25 @@ evaluation in TFX. TensorFlow Model Analysis allows you to perform model
evaluations in the TFX pipeline, and view resultant metrics and plots in a
Jupyter notebook. Specifically, it can provide:
-* [Metrics](../model_analysis/metrics) computed on entire training and holdout
+* [Metrics](https://www.tensorflow.org/tfx/model_analysis/metrics) computed on entire training and holdout
dataset, as well as next-day evaluations
* Tracking metrics over time
* Model quality performance on different feature slices
-* [Model validation](../model_analysis/model_validations) for ensuring that
+* [Model validation](https://www.tensorflow.org/tfx/model_analysis/model_validations) for ensuring that
model's maintain consistent performance
## Next Steps
-Try our [TFMA tutorial](../tutorials/model_analysis/tfma_basic).
+Try our [TFMA tutorial](https://www.tensorflow.org/tfx/tutorials/model_analysis/tfma_basic).
Check out our [github](https://github.com/tensorflow/model-analysis) page for
details on the supported
-[metrics and plots](../model_analysis/metrics) and associated notebook
-[visualizations](../model_analysis/visualizations).
+[metrics and plots](https://www.tensorflow.org/tfx/model_analysis/metrics) and associated notebook
+[visualizations](https://www.tensorflow.org/tfx/model_analysis/visualizations).
-See the [installation](../model_analysis/install) and
-[getting started](../model_analysis/get_started) guides for information and
-examples on how to get [set up](../model_analysis/setup) in a standalone
+See the [installation](https://www.tensorflow.org/tfx/model_analysis/install) and
+[getting started](https://www.tensorflow.org/tfx/model_analysis/get_started) guides for information and
+examples on how to get [set up](https://www.tensorflow.org/tfx/model_analysis/setup) in a standalone
pipeline. Recall that TFMA is also used within the [Evaluator](evaluator.md)
component in TFX, so these resources will be useful for getting started in TFX
as well.
diff --git a/docs/guide/tft_bestpractices.md b/docs/guide/tft_bestpractices.md
index 4beb024b59..11bd10ad52 100644
--- a/docs/guide/tft_bestpractices.md
+++ b/docs/guide/tft_bestpractices.md
@@ -22,7 +22,7 @@ and the TensorFlow
[Keras](https://www.tensorflow.org/guide/keras/overview) API.
The second document,
-[Data preprocessing for ML with Google Cloud](../tutorials/transform/data_preprocessing_with_cloud),
+[Data preprocessing for ML with Google Cloud](../../tutorials/transform/data_preprocessing_with_cloud),
provides a step-by-step tutorial for how to implement a `tf.Transform` pipeline.
## Introduction
@@ -100,7 +100,7 @@ meanings:
features that are created by performing certain ML-specific operations on
the columns in the prepared dataset, and creating new features for your
model during training and prediction, as described later in
- [Preprocessing operations](#preprocessing_operations).
+ [Preprocessing operations](#preprocessing-operations).
Examples of these operations include scaling numerical columns to a value
between 0 and 1, clipping values, and
[one-hot-encoding](https://developers.google.com/machine-learning/glossary/#one-hot_encoding){: .external }
@@ -109,12 +109,17 @@ meanings:
The following diagram, figure 1, shows the steps that are involved in preparing
preprocessed data:
-