Skip to content

Commit 5d747e2

Browse files
committed
Rename programmers_guide/ to guide/ in tf-models.
1 parent 7c5c014 commit 5d747e2

File tree

14 files changed

+33
-33
lines changed

14 files changed

+33
-33
lines changed

official/boosted_trees/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ We use Gradient Boosted Trees algorithm to distinguish the two classes.
99
The code sample uses the high level `tf.estimator.Estimator` and `tf.data.Dataset`. These APIs are great for fast iteration and quickly adapting models to your own datasets without major code overhauls. It allows you to move from single-worker training to distributed training, and makes it easy to export model binaries for prediction. Here, for further simplicity and faster execution, we use a utility function `tf.contrib.estimator.boosted_trees_classifier_train_in_memory`. This utility function is especially effective when the input is provided as in-memory data sets like numpy arrays.
1010

1111
An input function for the `Estimator` typically uses `tf.data.Dataset` API, which can handle various data control like streaming, batching, transform and shuffling. However `boosted_trees_classifier_train_in_memory()` utility function requires that the entire data is provided as a single batch (i.e. without using `batch()` API). Thus in this practice, simply `Dataset.from_tensors()` is used to convert numpy arrays into structured tensors, and `Dataset.zip()` is used to put features and label together.
12-
For further references of `Dataset`, [Read more here](https://www.tensorflow.org/programmers_guide/datasets).
12+
For further references of `Dataset`, [Read more here](https://www.tensorflow.org/guide/datasets).
1313

1414
## Running the code
1515
First make sure you've [added the models folder to your Python path](/official/#running-the-models); otherwise you may encounter an error like `ImportError: No module named official.boosted_trees`.
@@ -53,13 +53,13 @@ tensorboard --logdir=/tmp/higgs_model # set logdir as --model_dir set during tr
5353
```
5454

5555
## Inference with SavedModel
56-
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/programmers_guide/saved_model) format by using the argument `--export_dir`:
56+
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/guide/saved_model) format by using the argument `--export_dir`:
5757

5858
```
5959
python train_higgs.py --export_dir /tmp/higgs_boosted_trees_saved_model
6060
```
6161

62-
After the model finishes training, use [`saved_model_cli`](https://www.tensorflow.org/programmers_guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
62+
After the model finishes training, use [`saved_model_cli`](https://www.tensorflow.org/guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
6363

6464
Try the following commands to inspect the SavedModel:
6565

official/mnist/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ python mnist_test.py --benchmarks=.
3232

3333
## Exporting the model
3434

35-
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/programmers_guide/saved_model) format by using the argument `--export_dir`:
35+
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/guide/saved_model) format by using the argument `--export_dir`:
3636

3737
```
3838
python mnist.py --export_dir /tmp/mnist_saved_model
@@ -41,7 +41,7 @@ python mnist.py --export_dir /tmp/mnist_saved_model
4141
The SavedModel will be saved in a timestamped directory under `/tmp/mnist_saved_model/` (e.g. `/tmp/mnist_saved_model/1513630966/`).
4242

4343
**Getting predictions with SavedModel**
44-
Use [`saved_model_cli`](https://www.tensorflow.org/programmers_guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
44+
Use [`saved_model_cli`](https://www.tensorflow.org/guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
4545

4646
```
4747
saved_model_cli run --dir /tmp/mnist_saved_model/TIMESTAMP --tag_set serve --signature_def classify --inputs image=examples.npy

official/transformer/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -214,7 +214,7 @@ big | 28.9
214214
demonstration purposes only, but will be optimized in the coming weeks.
215215

216216
## Export trained model
217-
To export the model as a Tensorflow [SavedModel](https://www.tensorflow.org/programmers_guide/saved_model) format, use the argument `--export_dir` when running `transformer_main.py`. A folder will be created in the directory with the name as the timestamp (e.g. $EXPORT_DIR/1526427396).
217+
To export the model as a Tensorflow [SavedModel](https://www.tensorflow.org/guide/saved_model) format, use the argument `--export_dir` when running `transformer_main.py`. A folder will be created in the directory with the name as the timestamp (e.g. $EXPORT_DIR/1526427396).
218218

219219
```
220220
EXPORT_DIR=$HOME/transformer/saved_model
@@ -366,4 +366,4 @@ The [newstest2014 files](test_data) are extracted from the [NMT Seq2Seq tutorial
366366

367367
Example: Consider a training a dataset with 100 examples that is divided into 20 batches with 5 examples per batch. A single training step trains the model on one batch. After 20 training steps, the model will have trained on every batch in the dataset, or one epoch.
368368

369-
**Subtoken**: Words are referred to as tokens, and parts of words are referred to as 'subtokens'. For example, the word 'inclined' may be split into `['incline', 'd_']`. The '\_' indicates the end of the token. The subtoken vocabulary list is guaranteed to contain the alphabet (including numbers and special characters), so all words can be tokenized.
369+
**Subtoken**: Words are referred to as tokens, and parts of words are referred to as 'subtokens'. For example, the word 'inclined' may be split into `['incline', 'd_']`. The '\_' indicates the end of the token. The subtoken vocabulary list is guaranteed to contain the alphabet (including numbers and special characters), so all words can be tokenized.

official/wide_deep/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ For the purposes of this example code, the Census Income Data Set was chosen to
1010

1111
The code sample in this directory uses the high level `tf.estimator.Estimator` API. This API is great for fast iteration and quickly adapting models to your own datasets without major code overhauls. It allows you to move from single-worker training to distributed training, and it makes it easy to export model binaries for prediction.
1212

13-
The input function for the `Estimator` uses `tf.contrib.data.TextLineDataset`, which creates a `Dataset` object. The `Dataset` API makes it easy to apply transformations (map, batch, shuffle, etc.) to the data. [Read more here](https://www.tensorflow.org/programmers_guide/datasets).
13+
The input function for the `Estimator` uses `tf.contrib.data.TextLineDataset`, which creates a `Dataset` object. The `Dataset` API makes it easy to apply transformations (map, batch, shuffle, etc.) to the data. [Read more here](https://www.tensorflow.org/guide/datasets).
1414

1515
The `Estimator` and `Dataset` APIs are both highly encouraged for fast development and efficient training.
1616

@@ -48,13 +48,13 @@ tensorboard --logdir=/tmp/census_model
4848
```
4949

5050
## Inference with SavedModel
51-
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/programmers_guide/saved_model) format by using the argument `--export_dir`:
51+
You can export the model into Tensorflow [SavedModel](https://www.tensorflow.org/guide/saved_model) format by using the argument `--export_dir`:
5252

5353
```
5454
python wide_deep.py --export_dir /tmp/wide_deep_saved_model
5555
```
5656

57-
After the model finishes training, use [`saved_model_cli`](https://www.tensorflow.org/programmers_guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
57+
After the model finishes training, use [`saved_model_cli`](https://www.tensorflow.org/guide/saved_model#cli_to_inspect_and_execute_savedmodel) to inspect and execute the SavedModel.
5858

5959
Try the following commands to inspect the SavedModel:
6060

research/astronet/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@ the second deepest transits).
207207

208208
To train a model to identify exoplanets, you will need to provide TensorFlow
209209
with training data in
210-
[TFRecord](https://www.tensorflow.org/programmers_guide/datasets) format. The
210+
[TFRecord](https://www.tensorflow.org/guide/datasets) format. The
211211
TFRecord format consists of a set of sharded files containing serialized
212212
`tf.Example` [protocol buffers](https://developers.google.com/protocol-buffers/).
213213

@@ -343,7 +343,7 @@ bazel-bin/astronet/train \
343343
--model_dir=${MODEL_DIR}
344344
```
345345

346-
Optionally, you can also run a [TensorBoard](https://www.tensorflow.org/programmers_guide/summaries_and_tensorboard)
346+
Optionally, you can also run a [TensorBoard](https://www.tensorflow.org/guide/summaries_and_tensorboard)
347347
server in a separate process for real-time
348348
monitoring of training progress and evaluation metrics.
349349

research/seq2species/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -117,8 +117,8 @@ python seq2species/run_training.py --train_files ${TFRECORD}
117117
--logdir $HOME/seq2species
118118
```
119119
This will output [TensorBoard
120-
summaries](https://www.tensorflow.org/programmers_guide/summaries_and_tensorboard), [TensorFlow
121-
checkpoints](https://www.tensorflow.org/programmers_guide/variables#checkpoint_files), Seq2LabelModelInfo and
120+
summaries](https://www.tensorflow.org/guide/summaries_and_tensorboard), [TensorFlow
121+
checkpoints](https://www.tensorflow.org/guide/variables#checkpoint_files), Seq2LabelModelInfo and
122122
Seq2LabelExperimentMeasures metadata to the logdir `$HOME/seq2species`.
123123

124124
### Preprocessed Seq2Species Data

research/tcn/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ it to the TFRecord format expected by this library.
9292
## Data Pipelines
9393

9494
We use the [tf.data.Dataset
95-
API](https://www.tensorflow.org/programmers_guide/datasets) to construct input
95+
API](https://www.tensorflow.org/guide/datasets) to construct input
9696
pipelines that feed training, evaluation, and visualization. These pipelines are
9797
defined in `data_providers.py`.
9898

samples/core/get_started/basic_classification.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -129,7 +129,7 @@
129129
"source": [
130130
"In this guide, we will train a neural network model to classify images of clothing, like sneakers and shirts. It's fine if you don't understand all the details, this is a fast-paced overview of a complete TensorFlow program with the details explained as we go.\n",
131131
"\n",
132-
"This guide uses [tf.keras](https://www.tensorflow.org/programmers_guide/keras), a high-level API to build and train models in TensorFlow."
132+
"This guide uses [tf.keras](https://www.tensorflow.org/guide/keras), a high-level API to build and train models in TensorFlow."
133133
]
134134
},
135135
{

samples/core/get_started/basic_regression.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -120,7 +120,7 @@
120120
"\n",
121121
"This notebook builds a model to predict the median price of homes in a Boston suburb during the mid-1970s. To do this, we'll provide the model with some data points about the suburb, such as the crime rate and the local property tax rate.\n",
122122
"\n",
123-
"This example uses the `tf.keras` API, see [this guide](https://www.tensorflow.org/programmers_guide/keras) for details."
123+
"This example uses the `tf.keras` API, see [this guide](https://www.tensorflow.org/guide/keras) for details."
124124
]
125125
},
126126
{

samples/core/get_started/basic_text_classification.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -118,7 +118,7 @@
118118
"\n",
119119
"We'll use the [IMDB dataset](https://www.tensorflow.org/api_docs/python/tf/keras/datasets/imdb) that contains the text of 50,000 movie reviews from the [Internet Movie Database](https://www.imdb.com/). These are split into 25,000 reviews for training and 25,000 reviews for testing. The training and testing sets are *balanced*, meaning they contain an equal number of positive and negative reviews. \n",
120120
"\n",
121-
"This notebook uses [tf.keras](https://www.tensorflow.org/programmers_guide/keras), a high-level API to build and train models in TensorFlow."
121+
"This notebook uses [tf.keras](https://www.tensorflow.org/guide/keras), a high-level API to build and train models in TensorFlow."
122122
]
123123
},
124124
{

0 commit comments

Comments
 (0)