vit-large-patch16-384Model41/100 via “feature extraction and embedding generation for downstream tasks”
image-classification model by undefined. 4,74,363 downloads.
Unique: Extracts 1024-dimensional embeddings from the transformer's [CLS] token (global image representation) after 24 layers of multi-head self-attention, capturing long-range dependencies across all image patches. Unlike CNN-based feature extractors (ResNet) that produce spatial feature maps, ViT embeddings are fully global and normalized, making them directly suitable for vector similarity search without additional pooling or normalization steps.
vs others: Produces more semantically meaningful embeddings than ResNet features for fine-grained visual similarity due to global receptive field; embeddings are directly comparable across images without spatial alignment, enabling efficient nearest-neighbor search; requires more computational resources for embedding generation than lightweight CNN models