Skip to content

Commit 7016919

Browse files
authored
Update word-level LM arguments in README (pytorch#786)
1 parent 59423c5 commit 7016919

File tree

1 file changed

+22
-23
lines changed

1 file changed

+22
-23
lines changed

word_language_model/README.md

+22-23
Original file line numberDiff line numberDiff line change
@@ -25,29 +25,28 @@ The `main.py` script accepts the following arguments:
2525

2626
```bash
2727
optional arguments:
28-
-h, --help show this help message and exit
29-
--data DATA location of the data corpus
30-
--model MODEL type of recurrent net (RNN_TANH, RNN_RELU, LSTM, GRU)
31-
--emsize EMSIZE size of word embeddings
32-
--nhid NHID number of hidden units per layer
33-
--nlayers NLAYERS number of layers
34-
--lr LR initial learning rate
35-
--clip CLIP gradient clipping
36-
--epochs EPOCHS upper epoch limit
37-
--batch_size N batch size
38-
--bptt BPTT sequence length
39-
--dropout DROPOUT dropout applied to layers (0 = no dropout)
40-
--decay DECAY learning rate decay per epoch
41-
--tied tie the word embedding and softmax weights
42-
--seed SEED random seed
43-
--cuda use CUDA
44-
--log-interval N report interval
45-
--save SAVE path to save the final model
46-
--onnx-export path to export the final model in onnx format
47-
--transformer_head N the number of heads in the encoder/decoder of the transformer model
48-
--transformer_encoder_layers N the number of layers in the encoder of the transformer model
49-
--transformer_decoder_layers N the number of layers in the decoder of the transformer model
50-
--transformer_d_ff N the number of nodes on the hidden layer in feed forward nn
28+
-h, --help show this help message and exit
29+
--data DATA location of the data corpus
30+
--model MODEL type of recurrent net (RNN_TANH, RNN_RELU, LSTM, GRU,
31+
Transformer)
32+
--emsize EMSIZE size of word embeddings
33+
--nhid NHID number of hidden units per layer
34+
--nlayers NLAYERS number of layers
35+
--lr LR initial learning rate
36+
--clip CLIP gradient clipping
37+
--epochs EPOCHS upper epoch limit
38+
--batch_size N batch size
39+
--bptt BPTT sequence length
40+
--dropout DROPOUT dropout applied to layers (0 = no dropout)
41+
--tied tie the word embedding and softmax weights
42+
--seed SEED random seed
43+
--cuda use CUDA
44+
--log-interval N report interval
45+
--save SAVE path to save the final model
46+
--onnx-export ONNX_EXPORT
47+
path to export the final model in onnx format
48+
--nhead NHEAD the number of heads in the encoder/decoder of the
49+
transformer model
5150
```
5251

5352
With these arguments, a variety of models can be tested.

0 commit comments

Comments
 (0)