metadata
library_name: transformers
license: apple-ascl
metrics:
- accuracy
model-index:
- name: aimv2-3B-patch14-224
results:
- dataset:
name: imagenet-1k
type: imagenet-1k
metrics:
- name: Accuracy
type: accuracy
value: 88.5
verified: false
task:
name: Classification
type: classification
- dataset:
name: inaturalist-18
type: inaturalist-18
metrics:
- name: Accuracy
type: accuracy
value: 81.5
verified: false
task:
name: Classification
type: classification
- dataset:
name: cifar10
type: cifar10
metrics:
- name: Accuracy
type: accuracy
value: 99.5
verified: false
task:
name: Classification
type: classification
- dataset:
name: cifar100
type: cifar100
metrics:
- name: Accuracy
type: accuracy
value: 94.3
verified: false
task:
name: Classification
type: classification
- dataset:
name: food101
type: food101
metrics:
- name: Accuracy
type: accuracy
value: 96.8
verified: false
task:
name: Classification
type: classification
- dataset:
name: dtd
type: dtd
metrics:
- name: Accuracy
type: accuracy
value: 88.9
verified: false
task:
name: Classification
type: classification
- dataset:
name: oxford-pets
type: oxford-pets
metrics:
- name: Accuracy
type: accuracy
value: 97.1
verified: false
task:
name: Classification
type: classification
- dataset:
name: stanford-cars
type: stanford-cars
metrics:
- name: Accuracy
type: accuracy
value: 96.5
verified: false
task:
name: Classification
type: classification
- dataset:
name: camelyon17
type: camelyon17
metrics:
- name: Accuracy
type: accuracy
value: 93.5
verified: false
task:
name: Classification
type: classification
- dataset:
name: patch-camelyon
type: patch-camelyon
metrics:
- name: Accuracy
type: accuracy
value: 89.4
verified: false
task:
name: Classification
type: classification
- dataset:
name: rxrx1
type: rxrx1
metrics:
- name: Accuracy
type: accuracy
value: 7.3
verified: false
task:
name: Classification
type: classification
- dataset:
name: eurosat
type: eurosat
metrics:
- name: Accuracy
type: accuracy
value: 99
verified: false
task:
name: Classification
type: classification
- dataset:
name: fmow
type: fmow
metrics:
- name: Accuracy
type: accuracy
value: 64.2
verified: false
task:
name: Classification
type: classification
- dataset:
name: domainnet-infographic
type: domainnet-infographic
metrics:
- name: Accuracy
type: accuracy
value: 72.2
verified: false
task:
name: Classification
type: classification
pipeline_tag: image-feature-extraction
tags:
- vision
- image-feature-extraction
- mlx
- pytorch
Introduction
[AIMv2 Paper
] [BibTeX
]
We introduce the AIMv2 family of vision models pre-trained with a multimodal autoregressive objective. AIMv2 pre-training is simple and straightforward to train and scale effectively. Some AIMv2 highlights include:
- Outperforms OAI CLIP and SigLIP on the majority of multimodal understanding benchmarks.
- Outperforms DINOv2 on open-vocabulary object detection and referring expression comprehension.
- Exhibits strong recognition performance with AIMv2-3B achieving 89.5% on ImageNet using a frozen trunk.
Usage
PyTorch
import requests
from PIL import Image
from transformers import AutoImageProcessor, AutoModel
url = "http://images.cocodataset.org/val2017/000000039769.jpg"
image = Image.open(requests.get(url, stream=True).raw)
processor = AutoImageProcessor.from_pretrained(
"apple/aimv2-3B-patch14-224",
)
model = AutoModel.from_pretrained(
"apple/aimv2-3B-patch14-224",
trust_remote_code=True,
)
inputs = processor(images=image, return_tensors="pt")
outputs = model(**inputs)
JAX
import requests
from PIL import Image
from transformers import AutoImageProcessor, FlaxAutoModel
url = "http://images.cocodataset.org/val2017/000000039769.jpg"
image = Image.open(requests.get(url, stream=True).raw)
processor = AutoImageProcessor.from_pretrained(
"apple/aimv2-3B-patch14-224",
)
model = FlaxAutoModel.from_pretrained(
"apple/aimv2-3B-patch14-224",
trust_remote_code=True,
)
inputs = processor(images=image, return_tensors="jax")
outputs = model(**inputs)
Citation
If you find our work useful, please consider citing us as:
@misc{fini2024multimodal,
title = {Multimodal Autoregressive Pre-training of Large Vision Encoders},
author = {Enrico Fini and Mustafa Shukor and Xiujun Li and Philipp Dufter and Michal Klein and David Haldimann and Sai Aitharaju and Victor Guilherme Turrisi da Costa and Louis Béthune and Zhe Gan and Alexander T Toshev and Marcin Eichner and Moin Nabi and Yinfei Yang and Joshua M. Susskind and Alaaeldin El-Nouby},
year = {2024},
archivePrefix = {arXiv},
primaryClass = {cs.CV},
}