Use Grounded EdgeSAM to train a Instance Segmentation model

Before you can train a computer vision model, you need labeled data on which to train your model. The more accurate the labels, or annotations, are, the higher the performance the model will achieve.

Overview

Autodistill allows you to use state-of-the-art foundation models that know a lot about a variety of objects to label data for your project. You can then train a new model with your labeled data. This whole process uses around a dozen lines of code.

To learn more about how autodistill works, read our overview guide or watch our YouTube tutorial.

In this guide, we will show you how to use

Grounded EdgeSAM

to train a

model.

To label data for a model with Grounded EdgeSAM, you will:

  1. Install Autodistill
  2. Create a dataset
  3. Build a prompt to label images in the dataset
  4. Label our data on Autodistill
  5. Train a new model
  6. (Optional) Train a model or export your data

Let's get started!

Install Dependencies

First, install Autodistill and the required model dependencies:

pip install autodistill autodistill-grounded-edgesam

Create a Dataset

Before you can label a dataset, you need a dataset with which to work.

Roboflow has a few resources that can help you create a dataset for your project:

You can use any folder of images you have on your local machine with Autodistill, too.

Label Images

Autodistill has two model types:

  1. A Base Model, which automatically labels your data, and;
  2. A Target Model, which trains on your labeled data.

To label your dataset with a Base Model, you need to provide prompt(s) that are relevant to the classes you want to label.

Replace "example" below with the prompt you want to use. Replace "class" with the name of the class you want the prompt results to be saved as in your dataset. Also, replace the IMAGE_NAME with an image from your dataset.

The code cell below loads the base model with your prompt on the provided image, then visualizes the results.

You may need to experiment with a few prompts.


from autodistill_grounded_edgesam import GroundedEdgeSAM
from autodistill.detection import CaptionOntology
from autodistill.utils import plot
import cv2

# define an ontology to map class names to our GroundedEdgeSAM prompt
# the ontology dictionary has the format {caption: class}
# where caption is the prompt sent to the base model, and class is the label that will
# be saved for that caption in the generated annotations
# then, load the model
base_model = GroundedEdgeSAM(
    ontology=CaptionOntology(
        {
            "person": "person",
            "forklift": "forklift",
        }
    )
)

# run inference on a single image
results = base_model.predict("logistics.jpeg")

plot(
    image=cv2.imread("logistics.jpeg"),
    classes=base_model.ontology.classes(),
    detections=results
)

# label a folder of images
base_model.label("./context_images", extension=".jpeg")

To start labeling your images, run the following code:

base_model.label(input_folder="./images", output_folder="./dataset")

Train a Model

To train a

model using your newly-labeled dataset, run the following code:

After running this cell, you will have model weights that you can use to run inference on your new model.

Upload Model to Roboflow (Optional)

You can deploy your trained model to Roboflow. By deploying your model to Roboflow, you can run inference on our infinitely-scalable API. As your inference demands grow, our you will continue to see high levels of performance thanks to autoscaling infrastructure that is always on.

Roboflow offers:

Fully managed, infinitely-scalable APIs for deployed models

SDKs for common deployment targets (NVIDIA Jetson, Luxonis OAK, Docker, and more)

SOC II Type 1 Compliant

Trusted by 250,000+ developers

To deploy your model to Roboflow, run the following code: