Roboflow Docs
DashboardForum
  • Build Vision Models with Roboflow
  • Quickstart
  • Roboflow Enterprise
  • Workspaces
    • Create a Workspace
    • Delete a Workspace
    • Add Team Members
    • Role-Based Access Control
  • Usage Based Pricing
  • Workflows
    • Create a Workflow
    • Build a Workflow
    • Test a Workflow
    • Deploy a Workflow
    • Workflow Examples
      • Multimodal Model Workflow
    • Share a Workflow
      • Workflow Sharing Configuration
    • Advance Workflow Topics
      • JSON Editor
  • Datasets
    • Create a Project
    • Upload Data
      • Import Data from Cloud Providers
        • AWS S3 Bucket
        • Azure Blob Storage
        • Google Cloud Storage
      • Upload Video
      • Import from Roboflow Universe
    • Manage Batches
    • Search a Dataset
    • Create a Dataset Version
    • Preprocess Images
    • Create Augmented Images
    • Add Tags to Images
    • Manage Classes
    • Edit Keypoint Skeletons
    • Create an Annotation Attribute
    • Export Versions
    • Dataset Analytics
    • Merge Projects
    • Delete an Image
    • Delete a Version
    • Delete a Project
    • Project Folders
  • Annotate
    • Annotation Tools
    • Use Roboflow Annotate
      • Annotate Keypoints
      • Label Assist (AI Labeling)
      • Enhanced Smart Polygon with SAM (AI Labeling)
      • Smart Polygon (AI Labeling)
      • Keyboard Shortcuts
      • Comment on an Image
      • Annotation History
      • Similarity Search
      • Box Prompting (AI Labeling)
    • Automated Annotation with Auto Label
    • Collaborate on Annotations
    • Annotation Insights
    • Labeling Best Practices
  • Train
    • Train a Model in Roboflow
      • Train from Scratch
      • Train from a Universe Checkpoint
      • Python Package
      • Roboflow Notebooks (GitHub)
    • Train from Azure Vision
    • Train from Google Cloud
    • View Training Results
    • Evaluate Trained Models
    • Custom Training Notebooks
  • Deploy
    • Deployment Overview
      • Roboflow Managed Deployments Overview
    • Serverless Hosted API
      • Object Detection
      • Classification
      • Instance Segmentation
        • Semantic Segmentation
      • Keypoint Detection
      • Foundation Models
        • CLIP
        • OCR
        • YOLO-World
      • Video Inference
        • Use a Fine-Tuned Model
        • Use CLIP
        • Use Gaze Detection
        • API Reference
        • Video Inference JSON Output Format
      • Pre-Trained Model APIs
        • Blur People API
        • OCR API
        • Logistics API
        • Image Tagging API
        • People Detection API
        • Fish Detection API
        • Bird Detection API
        • PPE Detection API
        • Barcode Detection API
        • License Plate Detection API
        • Ceramic Defect Detection API
        • Metal Defect Detection API
    • Serverless Hosted API V2
    • Dedicated Deployments
      • How to create a dedicated deployment (Roboflow App)
      • How to create a dedicated deployment (Roboflow CLI)
      • How to use a dedicated deployment
      • How to manage dedicated deployment using HTTP APIs
    • SDKs
      • Python inference-sdk
      • Web Browser
        • inferencejs Reference
        • inferencejs Requirements
      • Lens Studio
        • Changelog - Lens Studio
      • Mobile iOS
      • Luxonis OAK
    • Upload Custom Weights
    • Download Roboflow Model Weights
    • Enterprise Deployment
      • License Server
      • Offline Mode
      • Kubernetes
      • Docker Compose
    • Model Monitoring
      • Alerting
  • Roboflow CLI
    • Introduction
    • Installation and Authentication
    • Getting Help
    • Upload Dataset
    • Download Dataset
    • Run Inference
  • API Reference
    • Introduction
    • Python Package
    • REST API Structure
    • Authentication
    • Workspace and Project IDs
    • Workspaces
    • Workspace Image Query
    • Batches
    • Annotation Jobs
    • Projects
      • Initialize
      • Create
      • Project Folders API
    • Images
      • Upload Images
      • Image Details
      • Upload Dataset
      • Upload an Annotation
      • Search
      • Tags
    • Versions
      • View a Version
      • Create a Project Version
    • Inference
    • Export Data
    • Train a Model
    • Annotation Insights
      • Annotation Insights (Legacy Endpoint)
    • Model Monitoring
      • Custom Metadata
      • Inference Result Stats
  • Support
    • Share a Workspace with Support
    • Account Deletion
    • Frequently Asked Questions
Powered by GitBook
On this page
  • Serverless API
  • Workflows
  • Batch Processing
  • Dedicated Deployments

Was this helpful?

  1. Deploy
  2. Deployment Overview

Roboflow Managed Deployments Overview

PreviousDeployment OverviewNextServerless Hosted API

Last updated 1 month ago

Was this helpful?

Roboflow provides several managed deployment options that leverage our cloud infrastructure to run your models. These options are easy to use and offer excellent scalability options, making them ideal for a wide range of applications.

Serverless API

The allows you to run workflows and models directly on Roboflow's infrastructure through an infinitely-scalable API. This is the easiest way to deploy your models and get started with inference.

Benefits:

  • Scalability: The API automatically scales to handle your inference needs, so you don't have to worry about provisioning or managing servers.

  • Ease of Use: You can access your models through a simple REST API, making it easy to integrate inference into your applications.

  • No Infrastructure Management: Roboflow handles all the infrastructure, so you can focus on building your applications.

  • Workflow Support: All your workflows are available as API endpoints on the Serveless API, so you can easily run your workflows using a simple HTTP request

Limitations:

  • Warmup Requests: When you make requests that are going to require loading a model that hasn’t been loaded into any of the servers, initial requests may have increased latency of several seconds. After subsequent requests the request latency will drastically improve as your model gets cached in the currently running servers.

  • CPU based: The uses CPU for model inference; you may experience higher latency compared to dedicated deployments or self-hosted deployments and can’t use models that require a GPU (Serverless GPU API coming soon).

Workflows

The allows you to run in the Roboflow cloud. This enables you to build and run complex computer vision applications without managing your own infrastructure.

You can also run workflows on or self hosted inference servers, which enables you to use more powerful GPU based models and use .

You can learn more about how to create, test, and deploy Workflows .

Model Inference

  1. Obtain your API key from the Roboflow dashboard.

  2. Send a POST request to the API endpoint with your image and model information.

  3. Receive the inference results in JSON format.

Batch Processing

With configurable processing workflows, real-time monitoring, and event-based notifications, Roboflow Batch Processing helps you efficiently manage data processing, track progress, and integrate with other systems—making it easy to achieve your goals.

Benefits:

  • Scalability: The service automatically scales to your data volume, capable of processing millions of images and thousands of video files efficiently.

  • No Infrastructure Management: Roboflow handles all the infrastructure and data management, so you can focus on solving your business use-cases.

Limitations:

  • Asynchronous nature of processing: The Batch Processing service launches processing jobs to run in the background when compute resources are available. While it typically takes only a few minutes to provision the necessary servers, there is no guarantee of an exact start time for the job. As a result, the service is not suitable for real-time processing.

  • Custom Python Blocks not supported: Since the service runs on Roboflow's infrastructure, we currently do not support the execution of arbitrary code through Custom Python Blocks.

Dedicated Deployments

Benefits:

  • Consistent Performance: Dedicated resources ensure consistent performance for your models.

  • Resource Isolation: Your models run on isolated resources, preventing interference from other users.

  • GPU support: You can run large models that require GPU support on Dedicated Deployments (like e.g. SAM2, CogVML)

  • Custom Python Blocks: You can use custom python blocks in your workflows when deploying them on Dedicated Deployments.

Limitations:

  • Limited to US-Based Data Centers: Currently, Dedicated Deployments are only available in US-based data centers, which may result in higher latency for users in other regions.

In addition to you can also infer against a specific model using the . You can infer against any model you have trained on Robofolow, any of the , or find projects with trained model on

Overview for How to Use the :

See the docs for details and API specifications

is a fully managed solution powered by Workflows that allows you to process large volumes of videos and images without writing code. It offers an easy-to-use UI for quick tasks and a comprehensive API for automating data processing—fitting both small and large workloads.

Ease of Use: You can use the service in multiple ways—from a simple UI click to executing CLI commands, all the way to building that seamlessly integrate with your system.

provide dedicated GPUs and CPUs for running your models. This option offers consistent performance, resource isolation, and enhanced security, making it suitable for demanding applications and production workloads that require resource isolation or custom code execution.

workflows
Serverless Hosted API
supported foundation models
https://universe.roboflow.com
Serverless Hosted API
Serverless Hosted API
Roboflow Batch Processing
production-grade automations
Dedicated Deployments
Serverless Hosted API
Serverless Hosted API
Serverless Hosted API
Workflows
Custom Python Blocks
here
Dedicated Deployments