Skip to content
This repository was archived by the owner on Jul 7, 2023. It is now read-only.

Commit d48e7ae

Browse files
authored
Merge pull request #635 from stefan-it/doc-fixes
Minor documentation fixes
2 parents 7331b7c + 033b667 commit d48e7ae

File tree

4 files changed

+7
-7
lines changed

4 files changed

+7
-7
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,7 @@ For all translation problems, we suggest to try the Transformer model:
154154
this should reach a BLEU score of about 28 on the English-German data-set,
155155
which is close to state-of-the art. If training on a single GPU, try the
156156
`--hparams_set=transformer_base_single_gpu` setting. For very good results
157-
or larger data-sets (e.g., for English-French)m, try the big model
157+
or larger data-sets (e.g., for English-French), try the big model
158158
with `--hparams_set=transformer_big`.
159159

160160
## Basics

docs/distributed_training.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ training.
55

66
T2T uses TensorFlow Estimators and so distributed training is configured with
77
the `TF_CONFIG` environment variable that is read by the
8-
[RunConfig](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/learn/python/learn/estimators/run_config.py)
8+
[RunConfig](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/python/estimator/run_config.py)
99
along with a set of flags.
1010

1111
## `TF_CONFIG`

docs/index.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ For language modeling, we have these data-sets in T2T:
6969
* LM1B (a billion-word corpus): `--problems=languagemodel_lm1b32k` for
7070
subword-level modeling and `--problems=languagemodel_lm1b_characters`
7171
for character-level modeling.
72-
72+
7373
We suggest to start with `--model=transformer` on this task and use
7474
`--hparams_set=transformer_small` for PTB and
7575
`--hparams_set=transformer_base` for LM1B.
@@ -95,7 +95,7 @@ For speech-to-text, we have these data-sets in T2T:
9595
For summarizing longer text into shorter one we have these data-sets:
9696
* CNN/DailyMail articles summarized into a few sentences:
9797
`--problems=summarize_cnn_dailymail32k`
98-
98+
9999
We suggest to use `--model=transformer` and
100100
`--hparams_set=transformer_prepend` for this task.
101101
This yields good ROUGE scores.
@@ -118,5 +118,5 @@ For all translation problems, we suggest to try the Transformer model:
118118
this should reach a BLEU score of about 28 on the English-German data-set,
119119
which is close to state-of-the art. If training on a single GPU, try the
120120
`--hparams_set=transformer_base_single_gpu` setting. For very good results
121-
or larger data-sets (e.g., for English-French)m, try the big model
121+
or larger data-sets (e.g., for English-French), try the big model
122122
with `--hparams_set=transformer_big`.

tensor2tensor/models/lstm.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ def dropout_lstm_cell():
6565
attention_mechanism_class = tf.contrib.seq2seq.BahdanauAttention
6666
else:
6767
raise ValueError("Unknown hparams.attention_mechanism = %s, must be "
68-
"luong or bahdanu." % hparams.attention_mechanism)
68+
"luong or bahdanau." % hparams.attention_mechanism)
6969
attention_mechanism = attention_mechanism_class(
7070
hparams.hidden_size, encoder_outputs)
7171

@@ -338,7 +338,7 @@ def lstm_attention():
338338

339339
@registry.register_hparams
340340
def lstm_bahdanau_attention_multi():
341-
"""Multi-head Bahdanu attention."""
341+
"""Multi-head Bahdanau attention."""
342342
hparams = lstm_bahdanau_attention()
343343
hparams.num_heads = 4
344344
return hparams

0 commit comments

Comments
 (0)