license: apache-2.0 | |
Model: SatMAE (https://arxiv.org/abs/2207.08051) | |
Variant: vitlarge-fmow-pretrain-800 | |
Example Usage: | |
```python | |
from huggingface_hub import hf_hub_download | |
import torch | |
hf_hub_download("MVRL/satmae-vitlarge-fmow-pretrain-800", "model.py", local_dir=".") | |
from model import MaskedAutoencoderViT | |
model = MaskedAutoencoderViT.from_pretrained("MVRL/satmae-vitlarge-fmow-pretrain-800") | |
print(model.forward_encoder(torch.randn(1, 3, 224, 224), mask_ratio=0.0)[0].shape) | |
``` |