We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

🚀 New model available: DeepSeek-V3.1 🚀

sentence-transformers/

clip-ViT-B-32

The CLIP model maps text and images to a shared vector space, enabling various applications such as image search, zero-shot image classification, and image clustering. The model can be used easily after installation, and its performance is demonstrated through zero-shot ImageNet validation set accuracy scores. Multilingual versions of the model are also available for 50+ languages.

Public
$0.005 / Mtoken
77
sentence-transformers/clip-ViT-B-32 cover image

Input

inputs
You can add more items with the button on the right

You need to login to use this model

Login

Settings

The service tier used for processing the request. When set to 'priority', the request will be processed with higher priority. 3

whether to normalize the computed embeddings 2

The number of dimensions in the embedding. If not provided, the model's default will be used.If provided bigger than model's default, the embedding will be padded with zeros. (Default: empty, 32 ≤ dimensions ≤ 8192)

Output

[
  [
    0,
    0.5,
    1
  ],
  [
    1,
    0.5,
    0
  ]
]
Model Information