--- name: DFN5B-CLIP-ViT-H-14-378-SAFETENSORS base_model: laion/CLIP-ViT-H-14-laion2B-s32B-b79K license: apple-amlr pipeline_tag: zero-shot-image-classification tags: - clip - Apple - OpenAI size: - 3946981876 - 3.95 GB tasks: - contrastive image-text - vision language: en papers: - https://arxiv.org/abs/2309.17425 datasets: - CommonPool-12.8B license_link: LICENSE --- > [!IMPORTANT] > Original Model Link : [https://huggingface.co/apple/DFN5B-CLIP-ViT-H-14-378](https://huggingface.co/apple/DFN5B-CLIP-ViT-H-14-378) > ``` name: DFN5B-CLIP-ViT-H-14-378-SAFETENSORS base_model: laion/CLIP-ViT-H-14-laion2B-s32B-b79K license: apple-amlr pipeline_tag: zero-shot-image-classification tags: - clip - Apple - OpenAI size: - 3946981876 - 3.95 GB tasks: - contrastive image-text - vision language: en papers: - https://arxiv.org/abs/2309.17425 datasets: - CommonPool-12.8B license_link: LICENS ``` # DFN5B-CLIP-ViT-H-14-378-SAFETENSORS CLIP ViT-H/14 trained on DFN-5B Data Filtering Network filtered from a 43B uncurated pool from CommonPool-12.8B + 30B public images.