Use the Roboflow CLI to create and manage Batch Processing jobs.
By installing inference-cli you gain access to the inference rf-cloud command, which allows you to interact with Batch Processing and Data Staging — the core components of Roboflow Batch Processing.
Batch ID format: Must be lowercase, at most 64 characters, with only letters, digits, hyphens (-), and underscores (_).
Cloud Storage
If your data is already in cloud storage (S3, Google Cloud Storage, or Azure), you can process it directly without downloading files locally. Install cloud storage support first:
For images:
For videos:
The --bucket-path parameter supports:
S3: s3://bucket-name/path/
Google Cloud Storage: gs://bucket-name/path/
Azure Blob Storage: az://container-name/path/
You can include glob patterns to filter files:
s3://my-bucket/training-data/**/*.jpg — All JPG files recursively
gs://my-bucket/videos/2024-*/*.mp4 — MP4 files in 2024-* folders
az://container/images/*.png — PNG files in images folder
Your cloud storage credentials are used only locally by the CLI to generate presigned URLs. They are never uploaded to Roboflow servers.
Generated presigned URLs are valid for 24 hours. Ensure your batch processing job completes within this timeframe.
For large datasets, the system automatically splits images into chunks of 20,000 files each. Videos work best in batches under 1,000.