Workflows
Roboflow Instance Segmentation Model to MQTT Publisher

Connect Roboflow Instance Segmentation Model with MQTT Publisher

Build a multi-stage computer vision pipeline by connecting Roboflow Instance Segmentation Model with MQTT Publisher and deploy a production application in minutes.
16,000+ organizations build with Roboflow
Roboflow Instance Segmentation Model

Roboflow Instance Segmentation Model

Run an instance segmentation model.
Run inference on an instance segmentation model hosted on or uploaded to Roboflow. You can query any model that is private to your account, or any public model available on [Roboflow Universe](https://universe.roboflow.com). You will need to set your Roboflow API key in your Inference environment to use this block. To learn more about setting your Roboflow API key, [refer to the Inference documentation](https://inference.roboflow.com/quickstart/configure_api_key/).
MQTT Publisher

MQTT Publisher

Publish serialized data payloads to an MQTT broker, handling connection management and error reporting.
The `MQTTPublisherBlock` is a sink block designed to send data payloads to a specified MQTT broker, facilitating real-time data distribution in IoT and other distributed systems. This block leverages the MQTT protocol, a lightweight messaging system widely used for devices with limited processing power and network bandwidth.

Deploy Workflows with a Hosted API or on the Edge

Use workflows with Roboflow Instance Segmentation Model and MQTT Publisher in production

Explore Popular Combinations

This is some text inside of a div block.
Model
Sink

Roboflow Instance Segmentation Model to Webhook Publish

Build a computer vision workflow that connects Roboflow Instance Segmentation Model to Webhook Publish.
Model
Sink

YOLOv8 to MQTT Publisher

Build a computer vision workflow that connects YOLOv8 to MQTT Publisher.
Model
Sink

Roboflow Keypoint Detection Model to MQTT Publisher

Build a computer vision workflow that connects Roboflow Keypoint Detection Model to MQTT Publisher.
Model
Sink

Roboflow Instance Segmentation Model to Super Annotator

Build a computer vision workflow that connects Roboflow Instance Segmentation Model to Super Annotator.
Model
Notifications

Roboflow Instance Segmentation Model to SMS

Build a computer vision workflow that connects Roboflow Instance Segmentation Model to SMS.
Model
Sink

YOLOv8 Classification to MQTT Publisher

Build a computer vision workflow that connects YOLOv8 Classification to MQTT Publisher.

How to Build a Workflow

Learn how to use a low-code open source platform to simplify building and deploying vision AI applications.
1
cube icon

Choose a Block

Choose from 40+ pre-built blocks that let you use custom models, open source models, LLM APIs, pre-built logic, and external applications. Blocks can be models from OpenAI or Meta AI, applications like Google Sheets or Pager Duty, and logic like filtering or cropping.
2
connected blocks icon

Connect Blocks

Each block can receive inputs, execute code, and send outputs to the next block in your Workflow. You can use the drag-and-drop UI to configure connections and see the JSON definitions of what’s happening behind the scenes.
3
rocket icon

Deploy Workflows

You’ll receive an output of the final result from your Workflow and the format you want it delivered in, like JSON. Once your Workflow produces sufficient results, you can use the Workflow as a hosted API endpoint or self-host in your own cloud, on-prem, or at the edge.

Deploy Workflows at Scale

Roboflow powers millions of daily inferences for the world’s largest enterprises on-device and in the cloud
Deploy your Workflows directly on fully managed infrastructure through an infinitely-scalable API endpoint for high volume workloads
Run Workflows on-device, internet connection optional, without the headache of environment management, dependencies, and managing CUDA versions.
Isolate dependencies in your software by using the Python SDK or HTTP API to operate and maintain your Workflows separate from other logic within your codebase
Supported devices include ARM CPU, x86 CPU, NVIDIA GPU, and NVIDIA Jetson

Customize Your Pipeline

Connect models from OpenAI or Meta AI, applications like Slack or Pager Duty, and logic like filtering or cropping.
View All Blocks