Roboflow Docs
DashboardResourcesProducts
  • Product Documentation
  • Developer Reference
  • Changelog
  • Roboflow Documentation
  • Quickstart
  • Workspaces
    • Workspaces, Projects, and Models
    • Create a Workspace
    • Rename a Workspace
    • Delete a Workspace
  • Team Members
    • Invite a Team Member
    • Role-Based Access Control (RBAC)
    • Change a Team Member Role
    • Remove a Team Member
  • Single Sign On (SSO)
  • Workflows
    • What is Workflows?
    • Create a Workflow
    • Build a Workflow
    • Test a Workflow
    • Deploy a Workflow
    • Workflow Examples
      • Multimodal Model Workflow
    • Share a Workflow
    • Workflows AI Assistant
  • Enterprise Integrations
  • Workflow Blocks
    • Run a Model
      • Object Detection Model
      • Single-Label Classification Model
    • Visualize Predictions
      • Bounding Box Visualization
      • Label Visualization
      • Circle Visualization
      • Background Color Visualization
      • Classification Label Visualization
      • Crop Visualization
  • Dataset Management
    • Create a Project
    • Upload Images, Videos, and Annotations
      • Import Data from Cloud Providers
        • AWS S3 Bucket
        • Azure Blob Storage
        • Google Cloud Storage
      • Import from Roboflow Universe
    • Manage Datasets
      • Dataset Batches
      • Search a Dataset
      • Set Dataset Classes
      • Add Tags to Images
      • Create an Annotation Attribute
      • Download an Image
      • Delete an Image
    • Dataset Versions
      • Create a Dataset Version
      • Preprocess Images
      • Augment Images
      • Delete a Version
      • Export a Dataset Version
    • Dataset Analytics
    • Merge Projects
    • Rename a Project
    • Delete a Project
    • Project Folders
    • Make a Project Public
  • Annotate
    • Introduction to Roboflow Annotate
    • Annotate an Image
      • Keyboard Shortcuts
      • Comment on an Image
      • Annotation History
      • Similarity Search
    • AI Labeling
      • Label Assist
      • Enhanced Smart Polygon with SAM
        • Smart Polygon (Legacy)
      • Box Prompting
      • Auto Label
    • Set Keypoint Skeletons
    • Annotate Keypoints
    • Annotate Multimodal Data
    • Collaborate on Labeling
    • Annotation Insights
  • Managed Labeling
  • Train
    • Train a Model
      • Train from a Universe Checkpoint
      • Train from Azure Vision
      • Train from Google Cloud
    • Roboflow Instant
    • Cancel a Training Job
    • Stop Training Early
    • View Training Results
    • View Trained Models
    • Evaluate Trained Models
  • Download a Dataset Version
  • Deploy
    • Deploy a Model or Workflow
    • Managed Deployments
    • Serverless Hosted API V2
      • Use in a Workflow
      • Use with the REST API
      • Run an Instant Model
    • Serverless Hosted API
      • Object Detection
      • Classification
      • Instance Segmentation
        • Semantic Segmentation
      • Keypoint Detection
      • Foundation Models
        • CLIP
        • OCR
        • YOLO-World
      • Video Inference
        • Use a Fine-Tuned Model
        • Use CLIP
        • Use Gaze Detection
        • API Reference
        • Video Inference JSON Output Format
      • Pre-Trained Model APIs
        • Blur People API
        • OCR API
        • Logistics API
        • Image Tagging API
        • People Detection API
        • Fish Detection API
        • Bird Detection API
        • PPE Detection API
        • Barcode Detection API
        • License Plate Detection API
        • Ceramic Defect Detection API
        • Metal Defect Detection API
    • Dedicated Deployments
      • Create a Dedicated Deployment
      • Make Requests to a Dedicated Deployment
      • Manage Dedicated Deployments with an API
    • Batch Processing
    • SDKs
      • Python inference-sdk
      • Web Browser
        • inferencejs Reference
        • inferencejs Requirements
      • Lens Studio
        • Changelog - Lens Studio
      • Luxonis OAK
    • Upload Custom Model Weights
    • Download Model Weights
    • Enterprise Deployment
      • License Server
      • Offline Mode
      • Kubernetes
      • Docker Compose
    • Monitor Deployed Models
      • Alerting
  • Universe
    • What is Roboflow Universe?
    • Find a Dataset on Universe
    • Explore Images in a Universe Dataset
    • Fork a Universe Dataset
    • Find a Model on Universe
    • Download a Universe Dataset
  • Set a Project Description
  • View Project Analytics
  • Support
    • Share a Workspace with Support
    • Delete Your Roboflow Account
    • Apply for Academic Credits
  • Billing
    • Premium Trial
    • Credits
      • View Credit Usage
      • Enable or Disable Flex Billing
      • Purchase Prepaid Credits
    • Plans
      • Purchase a Plan
      • Cancel a Plan
      • Update Billing Details
      • Update Payment Method
      • View Invoices
Powered by GitBook
On this page

Was this helpful?

  1. Deploy

Deploy a Model or Workflow

Learn how to deploy models trained on or uploaded to Roboflow.

PreviousDownload a Dataset VersionNextManaged Deployments

Last updated 4 days ago

Was this helpful?

You can deploy any model trained on or uploaded to Roboflow, and any Workflow, with Roboflow's deployment offerings.

Our deployment offerings fit into two categories:

  • : These options leverage Roboflow's cloud infrastructure to run your models, eliminating the need for you to manage your own hardware or software.

  • : These options allow you to deploy models on your own hardware, providing greater control over your environment and resources.

The following table summarizes the key features, benefits, and limitations of each deployment option:

Deployment Option
Description
Benefits
Limitations

Run workflows and models directly on Roboflow's infrastructure through an infinitely-scalable API.

Scalable, easy to use, no infrastructure management.

Limited control over resources, potential for higher latency for demanding applications.

Dedicated GPUs and CPUs for running workflows and models.

Support for GPU models, Video Streaming, Custom Python Blocks.

Limited to US-based data centers. Not autoscaling like Serverless API

Batch Processing

Managed pool of servers processing your images and videos with selected workflow.

Fully managed solution offering high data throughput and cost efficiency, seamlessly scalable to your data volume with GPU support.

Non-real-time processing and no support for Custom Python Blocks.

Self-Hosted Deployments

Run Inference on your own hardware.

Full control over resources and environment, potential for lower latency.

Requires infrastructure management and expertise.

Serverless Hosted API V2

Run workflows and models directly on Roboflow's infrastructure on GPU hardware.

Support for GPU models.

Limited control over resources, potential for higher latency for demanding applications or during periods of high load

What is Inference?

In computer vision, inference refers to the process of using a trained model to analyze new images or videos and make predictions. For example, an object detection model might be used to identify and locate objects in a video stream, or a classification model might be used to categorize images based on their content.

is an open-source project that provides a powerful and flexible framework for deploying computer vision models and workflows. It is s the engine that powers most of Roboflows managed deployment services. You can also self host it or use it to deploy your vision workflows to edge devices. Roboflow Inference offers a range of features and capabilities, including:

  • Support for various model architectures and tasks, including object detection, classification, instance segmentation, and more.

  • Workflows, which lets you build computer vision applications by combining different models, pre-built logic, and external applications by choosing from hundreds of building Blocks.

  • Hardware acceleration for optimized performance on different devices, including CPUs, GPUs, and edge devices like NVIDIA Jetson.

  • Multiprocessing for efficient use of resources.

  • Video decoding for seamless processing of video streams.

  • HTTP interface, APIs and docker images to simplify deployment

  • Integration with Roboflow's hosted deployment options and the Roboflow platform.

What is a Workflow?

enable you to build complex computer vision applications by combining different models, pre-built logic, and external applications. They provide a visual, low-code environment for designing and deploying sophisticated computer vision pipelines.

With Workflows, you can:

  • Chain multiple models together to perform complex tasks.

  • Add custom logic and decision-making to your applications.

  • Integrate with external systems and APIs.

  • Track, count, time, measure, and visualize objects in images and videos.

Choosing the Right Deployment Option

The best deployment option for you depends on your specific needs and requirements. Consider the following factors when making your decision:

  • Scalability: If your application needs to handle varying levels of traffic or data volume, the serverless API offers excellent scalability for real-time use-cases - otherwise Batch Processing is suggested option.

  • Latency: If you need low latency or video processing, dedicated deployments or self-hosted deployments with powerful hardware might be the best choice.

  • GPUs: If you need to run models that require a GPU (e.g. SAM2, CogVML, etc) you need to use a Dedicated Deployment with GPU machine type or self hosted on hardware that has GPUs available. (Serverless GPU API coming soon)

  • Control: Self-hosted deployments provide the most control over your environment and resources.

  • Expertise: Self-hosted deployments require more technical expertise to set up and manage.

There is great guide on how to choose the best deployment method for your use case in the inference getting started guide at:

Managed Deployments
Self-Hosted Deployments
Roboflow Inference
Workflows
https://inference.roboflow.com/start/getting-started/
Serverless API
Dedicated Deployments