
The goal of CLIP_benchmark is to evaluate CLIP-like models on a standard set of datasets on different tasks such as zero-shot classification and zero-shot retrieval, and captioning. Features Support for zero-shot classification and zero-shot retrieval, linear probing, and captioning. Support for OpenCLIP pre-trained models, Japanese CLIP, and NLLB CLIP for general multilingual abilities. Support various datasets from torchvision, tensorflow datasets, and VTAB. Support for various multilingual datasets for classification and retrieval Support for compositionality tasks
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
