Search results
Page title matches
- | hugging-face-uri = nlpconnect/vit-gpt2-image-captioning git clone https://huggingface.co/nlpconnect/vit-gpt2-image-captioning38 KB (4,971 words) - 03:33, 23 May 2023
- | hugging-face-uri = google/vit-base-patch16-224 | related-to = vit, vision, AutoTrain Compatible37 KB (4,937 words) - 03:33, 23 May 2023
- | hugging-face-uri = openai/clip-vit-large-patch14 git clone https://huggingface.co/openai/clip-vit-large-patch1438 KB (4,973 words) - 03:27, 23 May 2023
- | hugging-face-uri = openai/clip-vit-base-patch32 git clone https://huggingface.co/openai/clip-vit-base-patch3238 KB (4,973 words) - 03:27, 23 May 2023
- | hugging-face-uri = openai/clip-vit-base-patch16 git clone https://huggingface.co/openai/clip-vit-base-patch1638 KB (4,971 words) - 03:27, 23 May 2023
- #REDIRECT [[ViT-GPT2-image-captioning model]]45 bytes (4 words) - 23:16, 1 February 2023
Page text matches
- #REDIRECT [[ViT-GPT2-image-captioning model]]45 bytes (4 words) - 23:16, 1 February 2023
- | hugging-face-uri = openai/clip-vit-large-patch14599 bytes (82 words) - 15:17, 4 May 2023
- | hugging-face-uri = nlpconnect/vit-gpt2-image-captioning git clone https://huggingface.co/nlpconnect/vit-gpt2-image-captioning38 KB (4,971 words) - 03:33, 23 May 2023
- | hugging-face-uri = openai/clip-vit-base-patch16 git clone https://huggingface.co/openai/clip-vit-base-patch1638 KB (4,971 words) - 03:27, 23 May 2023
- | hugging-face-uri = openai/clip-vit-large-patch14 git clone https://huggingface.co/openai/clip-vit-large-patch1438 KB (4,973 words) - 03:27, 23 May 2023
- | hugging-face-uri = openai/clip-vit-base-patch32 git clone https://huggingface.co/openai/clip-vit-base-patch3238 KB (4,973 words) - 03:27, 23 May 2023
- | hugging-face-uri = google/vit-base-patch16-224 | related-to = vit, vision, AutoTrain Compatible37 KB (4,937 words) - 03:33, 23 May 2023
- '''[[ViT]]''' - [[Vision Transformer]] - https://arxiv.org/abs/2010.119294 KB (550 words) - 09:53, 14 May 2023
- | related-to = vit, dino, vision | all-tags = Feature Extraction, PyTorch, Transformers, imagenet-1k, vit, dino, vision, arxiv:2104.14294, License: apache-2.038 KB (5,144 words) - 03:28, 23 May 2023
- | related-to = vit, dino, vision ...tags = Feature Extraction, PyTorch, TensorFlow, Transformers, imagenet-1k, vit, dino, vision, arxiv:2104.14294, License: apache-2.038 KB (5,146 words) - 03:28, 23 May 2023
- ...earch/vision_transformer GitHub] || [[Computer Vision]] || [[Google]] || [[ViT]] ([[Vision Transformer]]) ||20 KB (1,948 words) - 23:18, 5 February 2024
- #A [[ViT-H image encoder]] that runs once per image and outputs an [[image embedding9 KB (1,300 words) - 15:16, 9 April 2023
- ...2x512 images from a subset of the LAION-5B database. It uses a frozen CLIP ViT-L/14 text encoder to condition the model on text prompts." <ref name="”418 KB (2,517 words) - 22:04, 27 May 2023
- | '''[[ViT]]''' || || [[Vision Transformer]]34 KB (4,201 words) - 04:37, 2 August 2023