This repository was archived by the owner on Nov 1, 2024. It is now read-only.
- Notifications
You must be signed in to change notification settings - Fork41
Omnivore: A Single Model for Many Visual Modalities
License
NotificationsYou must be signed in to change notification settings
facebookresearch/omnivore
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This repository contains PyTorch pretrained models, inference examples for the following papers:
Omnivore A single vision model for many different visual modalities, CVPR 2022 [bib]
@inproceedings{girdhar2022omnivore, title={{Omnivore: A Single Model for Many Visual Modalities}}, author={Girdhar, Rohit and Singh, Mannat and Ravi, Nikhila and van der Maaten, Laurens and Joulin, Armand and Misra, Ishan}, booktitle={CVPR}, year={2022}}
OmniMAE Single Model Masked Pretraining on Images and Videos [bib]
@article{girdhar2022omnimae, title={OmniMAE: Single Model Masked Pretraining on Images and Videos}, author={Girdhar, Rohit and El-Nouby, Alaaeldin and Singh, Mannat and Alwala, Kalyan Vasudev and Joulin, Armand and Misra, Ishan}, journal={arXiv preprint arXiv:2206.08356}, year={2022}}
OmniVision Our training pipeline supporting the multi-modal vision research.[bib]
We welcome your pull requests! Please seeCONTRIBUTING andCODE_OF_CONDUCT for more information.
Omnivore is released under the CC-BY-NC 4.0 license. SeeLICENSE for additional details. However the Swin Transformer implementation is additionally licensed under the Apache 2.0 license (seeNOTICE for additional details).
About
Omnivore: A Single Model for Many Visual Modalities
Resources
License
Code of conduct
Security policy
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
No releases published
Packages0
No packages published
Contributors8
Uh oh!
There was an error while loading.Please reload this page.