Comment on page
Create Augmented Images
Create augmented images to improve model performance.
Image augmentation is a step where augmentations are applied to existing images in your dataset. This process can help improve the ability of your model to generalize and thus perform more effectively on unseen images.
Roboflow supports the following augmentations:
- 90 degree rotation
- Random rotation
- Random crop
- Random shear
- Random noise
- Cutout (paid plans only)
- Mosaic (paid plans only)
We recommend starting a project with no augmentations. This allows you to evaluate the quality of your raw dataset. If you add augmentations and your dataset doesn't perform as well as expected, you will not have a baseline to which you can compare model performance.
If your doesn't perform well without augmentations, you may need to investigate class balance, data representation, and dataset size. When you have a dataset on which you have successfully trained a model without augmentations, you can add augmentations to help improve model performance.
Doing your augmentations through in a version ("offline augmentation") rather than at the time of training has a few key benefits.
- 1.Model reproducibility is increased. With Roboflow, you have a copy of how each image was augmented. For example, may find your model performs better on bright images rather than dark images, so you should collect more low-light training data.
- 2.Training time is decreased. Augmentations are CPU-constrained operations. When you’re training on your GPU and conducting augmentations on-the-fly, your GPU is often waiting for your CPU to provide augmented data at each epoch. That adds up!
- 3.Training costs are decreased. Because augmentations are CPU-constrained operations, your expensive, rented GPU is often waiting to be fed images for training.
To add augmentations, go to the Versions tab associated with your project in the Roboflow dashboard. Then, click "Augmentations" to set up augmentations for your project.
You can select how many times you seek a given image to be augmented. For example, sliding to 3 means each of your images will receive 2 random augmentations based on the settings you select.
Example case: 3x augmentation --> 1 of the created images is created only with the preprocessing settings you have applied. The other 2 images receive augmentations, leaving you with 3 times the number of images for each source image.
Augmentations are chained together, with randomization for the augmentation settings, and values for each setting, applied to each augmented image. Any images that appear as duplicates during this process are filtered out of the created version.
For example, if you select “flip horizontally” and “salt and pepper noise,” a given image will randomly be reflected as a horizontal flip and receive random salt and pepper noise.
Below are the augmentations supported by Roboflow. The parameters you can customize are in bullet points.
Randomly flip (reflect) an image vertically or horizontally. Annotations are correctly mirrored.
Randomly rotate an image 90 degrees or 180 degrees.
- Clockwise: Rotates an image 90 degrees clockwise.
- Counter Clockwise: Rotates an image 90 degrees counter clockwise.
- Upside Down: Rotates an image 180 degrees (upside down).
- Degrees: Select the highest amount an image will be randomly rotated clockwise or counter clockwise.
- Percent: The percent area of the original image to drop. (e.g. The percentage area of the original image to keep. (e.g. a higher percentage contains a smaller amount of the original image.)
Note: annotations are affected. At present, our implementation drops any annotations that are completely out of frame. We crop any annotation that are partially out of frame to be in line with the edge of the image. For these kept annotations, we currently keep any amount of the original object detection area. We will soon provide the ability for you to select what percentage of annotation area you seek to maintain -- for example, imagine you only want to keep annotations that have at least 80% of the area of their original bounding box -- that will be supported.
- Horizontal: Select the highest amount an image will be randomly sheared across its x-axis.
- Vertical: Select the highest amount an image will be randomly sheared across its y-axis.
Adjust the gamma exposure of an image to be brighter or darker.
- Percent: Select the percent up to which an image will be randomly brightened or darkened. Up to 100 percent bright (completely white) or 100 percent dark (completely black).
- Pixels: Determines the amount of blur applied to an image (i.e. the kernel size of the blurring process; all kernel sizes are odd). 25 pixels is max blur.
- Percent: Selects the percent of an image’s pixels that are affected, up to 25 percent.
Bounding box level augmentation creates new training data by only altering the content of a source image’s bounding boxes. In doing so, developers have greater control over creating training data that is more suitable to their problem’s conditions.
- A 2019 paper from Google researchers introduces the idea of using bounding box only augmentation to create optimal data for their models. In this paper, researchers showed bounding box only modifications create systemic improvements, especially for models that were fit on small datasets.
Last modified 8d ago