fix typo
Browse files
README.md
CHANGED
@@ -129,7 +129,7 @@ Perplexity: [https://en.wikipedia.org/wiki/Perplexity](https://en.wikipedia.org/
|
|
129 |
#### Testing Data
|
130 |
|
131 |
The IMDB dataset from Stanford NLP comes pre-split into training and testing data of 25k reviews each. Our preprocessing, which included the chunking of concatenated, tokenized inputs
|
132 |
-
into chunks of 256 tokens, increased these respective splits by approximately ~5k records each. We apply a single masking function to the
|
133 |
|
134 |
### Results
|
135 |
|
|
|
129 |
#### Testing Data
|
130 |
|
131 |
The IMDB dataset from Stanford NLP comes pre-split into training and testing data of 25k reviews each. Our preprocessing, which included the chunking of concatenated, tokenized inputs
|
132 |
+
into chunks of 256 tokens, increased these respective splits by approximately ~5k records each. We apply a single masking function to the evaluation dataset before testing as mentioned above.
|
133 |
|
134 |
### Results
|
135 |
|