No items found.
Use the widget below to experiment with MobileCLIP. You can detect COCO classes such as people, vehicles, animals, household items.
MobileCLIP is an image embedding model developed by Apple and introduced in the "MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training" paper.
You can use MobileCLIP to calculate image embeddings. These embeddings can be used for:
MobileCLIP
is licensed under a
license.
You can use Roboflow Inference to deploy a
MobileCLIP
API on your hardware. You can deploy the model on CPU (i.e. Raspberry Pi, AI PCs) and GPU devices (i.e. NVIDIA Jetson, NVIDIA T4).
Below are instructions on how to deploy your own model API.
First, install Autodistill and Autodistill MobileCLIP:
pip install autodistill-grounded-sam autodistill-mobileclip
Then, run:
from autodistill_mobileclip import MobileCLIP
from autodistill.detection import CaptionOntology
# define an ontology to map class names to our MobileCLIP prompt
# the ontology dictionary has the format {caption: class}
# where caption is the prompt sent to the base model, and class is the label that will
# be saved for that caption in the generated annotations
# then, load the model
base_model = MobileCLIP(
ontology=CaptionOntology(
{
"person": "person",
"a forklift": "forklift"
}
)
)
result = base_model.predict("image.jpeg")
print(result)