--- dataset_info: features: - name: image dtype: image - name: caption dtype: string splits: - name: train num_bytes: 171055893.125 num_examples: 1087 download_size: 170841790 dataset_size: 171055893.125 language: - en task_categories: - text-to-image annotations_creators: - machine-generated size_categories: - n<1K --- # Disclaimer This was inspired from https://huggingface.co/datasets/lambdalabs/pokemon-blip-captions # Dataset Card for A subset of Vivian Maier's photographs BLIP captions The captions are generated with the [pre-trained BLIP model](https://github.com/salesforce/BLIP). For each row the dataset contains `image` and `caption` keys. `image` is a varying size PIL jpeg, and `caption` is the accompanying text caption. Only a train split is provided. [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)