The roboflow.com/credits page mentions that 1 credit corresponds to 500 seconds of inference time. A more accurate formula is the following:
if x-remote-processing-time header is set: credits = (100ms + x-remote-processing-time) / 500,000mselse: credits = max(x-processing-time, 100ms) / 500,000ms
Where x-processing-time and x-remote-processing-time are HTTP Response headers, in float format (seconds). See roboflow.com/pricing for credit pricing.
Model Inference
In the example below, we run inference on coco/39 model (RF-DETR Small, 560x560). In response headers we can find x-processing-time , which is 81ms. In this case, we'd have credits = max(81, 100) / 500,000 = 0.0002 credits , or 0.2 credits per 1000 images.
If you run the same request a 10 minutes later, it could happen that the model has been unloaded and needs to be loaded to the GPU again - a cold start. Model loading can take up to a few seconds, and is highly correlated with the delay between inferences.
Formula: credits = max(1106, 100)/500,000 = 0.0022 , or 2.2 credits per 1000 (cold start) images.
Workflow run
For Workflows, we split model inference from general Workflow processing. This means that Workflow itself will be executed on (cheaper) CPU-only machines, and only use GPU machines for model inference, resulting in a more cost-effective processing.
License plate recognition Workflow with 2x object detection model, dynamic cropping, multiple visualizations, and Gemini for OCR
Formula: credits = (100ms + 1054ms)/500,000 , so 0.0023 credits for processing, and some tiny amount for the Gemini API call (depending on token count, see roboflow.com/credits).