--- language: - tr dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 292607887 num_examples: 81708 - name: validation num_bytes: 32771771 num_examples: 9079 download_size: 170747517 dataset_size: 325379658 configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* --- `habanoz/news-tr-90k` is split into training and validation splits. Columns are merged into a single text document using following format: ```text # {x['Title']} ## Özet {x['Summary']} ## İçerik {x['Text']} ``` A tokenizer is trained on this dataset with an 8K token vocabulary. According to the tokenizer, this dataset contains 62748074 (62.7M) training tokens and 7015414 (7M) validation tokens.