Update README.md
Browse files
README.md
CHANGED
@@ -9,6 +9,6 @@ size_categories:
|
|
9 |
- 10B<n<100B
|
10 |
---
|
11 |
|
12 |
-
[Fineweb sample-100BT](https://huggingface.co/datasets/HuggingFaceFW/fineweb) tokenized with the [gpt2 tokenizer](https://huggingface.co/openai-community/gpt2).
|
13 |
|
14 |
For efficiency, tokens are stored in uint16 numpy arrays. Shards are stored in the webdataset format.
|
|
|
9 |
- 10B<n<100B
|
10 |
---
|
11 |
|
12 |
+
[Fineweb sample-100BT](https://huggingface.co/datasets/HuggingFaceFW/fineweb) tokenized with the [gpt2 tokenizer](https://huggingface.co/openai-community/gpt2). Sequences are truncated to a length of 1024 tokens.
|
13 |
|
14 |
For efficiency, tokens are stored in uint16 numpy arrays. Shards are stored in the webdataset format.
|