🚀 New model available: DeepSeek-V3.1 🚀
openai/
The CLIP model was developed by OpenAI to investigate the robustness of computer vision models. It uses a Vision Transformer architecture and was trained on a large dataset of image-caption pairs. The model shows promise in various computer vision tasks but also has limitations, including difficulties with fine-grained classification and potential biases in certain applications.
Please upload an image file
You need to login to use this model
LoginSettings
dog (0.90)
cat (0.10)
© 2025 Deep Infra. All rights reserved.