Skip to content

Commit 2ce292b

Browse files
committed
typos
1 parent 3388d7e commit 2ce292b

File tree

1 file changed

+6
-5
lines changed

1 file changed

+6
-5
lines changed

README.md

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -52,15 +52,16 @@ As with the [TensorFlow code](https://github.com/openai/finetune-transformer-lm)
5252
python train.py --dataset rocstories --desc rocstories --submit --analysis --data_dir [path to data here]
5353
```
5454

55-
#### Accuracy on the ROCStories test set
55+
#### First experiments on the ROCStories test set
5656
Finetuning the PyTorch model for 3 Epochs on ROCStories takes 10 minutes to run on a single NVidia K-80.
5757

58-
The test accuracy of this PyTorch version (with the default TensorFlow hyper-parameters not finetuned for the differences between PyTorch and TensorFlow internal operations) is 83.43%.
58+
The test accuracy of this PyTorch version is 83.43% (with the default TensorFlow hyper-parameters not finetuned on the PyTorch model to take into account the differences between PyTorch and TensorFlow internals).
59+
5960
The authors reports a median accuracy with the TensorFlow code of 85.8%.
60-
The paper reports a best accuracy of 86.5%.
61-
The previous SOTA on the ROCStories dataset is 77.6 (Hidden Coherence Model of Chaturvedi et al. in "Story Comprehension for Predicting What Happens Next" EMNLP 2017. Which is a very nice paper by the way, you should check it out)
6261

63-
As noted by the author, the code can be non-deterministic due to various GPU ops.
62+
The paper reports a best single run accuracy of 86.5%.
63+
64+
The previous SOTA on the ROCStories dataset is 77.6% ("Hidden Coherence Model" of Chaturvedi et al. published in "Story Comprehension for Predicting What Happens Next" EMNLP 2017. Which is a very nice paper by the way, you should check it out)
6465

6566
### TO-DO list
6667
- [ ] Add Multi-GPU training logic

0 commit comments

Comments
 (0)