--- dataset_info: features: - name: image dtype: image - name: prompt dtype: string - name: label dtype: string splits: - name: train num_bytes: 23257283614.36 num_examples: 153128 download_size: 23241036646 dataset_size: 23257283614.36 configs: - config_name: default data_files: - split: train path: data/train-* task_categories: - object-detection language: - tr --- This dataset is combined and deduplicated version of [coco-2014](https://huggingface.co/datasets/detection-datasets/coco) and [coco-2017](https://huggingface.co/datasets/rafaelpadilla/coco2017) datasets for object detection. The labels are in Turkish and the dataset is in an instruction-tuning format with separate columns for prompts and completion labels. For the bounding boxes, a similar annotation scheme to that of [PaliGemma](https://huggingface.co/blog/paligemma#Detection) annotation is used. That is, ``` The bounding box coordinates are in the form of special tokens, where value is a number that represents a normalized coordinate. Each detection is represented by four location coordinates in the order x_min(left), y_min(top), x_max(right), y_max(bottom), followed by the label that was detected in that box. To convert values to coordinates, you first need to divide the numbers by 1024, then multiply y by the image height and x by its width. This will give you the coordinates of the bounding boxes, relative to the original image size. ```