Skip to content

Releases: naver-ai/vidt

ViDT+ Optimized

05 Apr 15:28
3c1d59a
Compare
Choose a tag to compare

ViDT+ models were trained for 150 epochs and with full proposed components.

ViDT+ models

05 Apr 14:58
ce0df48
Compare
Choose a tag to compare

We trained ViDT+ models for 50 epochs.

ViDT models trained with distillation

05 Nov 13:22
43ab7da
Compare
Choose a tag to compare

We trained ViDT models with distillation (token matching) for 50 epochs

ViDT models trained for 50 and 150 epochs

05 Nov 09:05
12b1593
Compare
Choose a tag to compare

There are ViDT pre-trained models for 50 and 150 epochs with different model sizes (from nano to base).
We activated auxiliary decoding loss and iterative box refinement.

Swin-nano pre-trained on ImageNet-1K

05 Nov 06:10
Compare
Choose a tag to compare

This is a pre-trained model called Swin-nano. The accuracy was 74.9% when trained for 300 epochs.