trainPoseMaskRCNN
Syntax
Description
trains a Pose Mask R-CNN network to perform six-degrees-of-freedom (6-DoF) pose estimation
for multiple object classes.net
= trainPoseMaskRCNN(trainingData
,network
,trainingMode
,options
)
Note
This functionality requires Deep Learning Toolbox™ and the Computer Vision Toolbox™ Model for Pose Mask R-CNN 6-DoF Object Pose Estimation. To use this functionality in parallel, you must have a license for Parallel Computing Toolbox™ and a CUDA® enabled NVIDIA® GPU. For information about the supported compute capabilities, see GPU Computing Requirements (Parallel Computing Toolbox).
[
also returns information on the training progress, such as training loss and accuracy, for
each iteration.net
,info
] = trainPoseMaskRCNN(trainingData
,network
,trainingMode
,options
)
[___] = trainPoseMaskRCNN(___,
specifies options using name-value arguments in addition to any combination of arguments
from previous syntaxes. For example, Name=Value
)NumRegionsToSample=64
specifies for
the trainPoseMaskRCNN
function to sample 64 region proposals from each
training image.
Input Arguments
Output Arguments
Tips
The
trainPoseMaskRCNN
function has a high GPU memory requirement. It is recommended to train a Pose Mask R-CNN network with at least 12 GB of available memory.To reduce memory consumption during training, you can decrease the value of the
NumRegionsToSample
name-value argument to limit the number of proposals from the region proposal stage. Note that this also reduces accuracy and increases convergence time.
Version History
Introduced in R2024a
See Also
posemaskrcnn
| predictPose
| trainingOptions
(Deep Learning Toolbox) | trainMaskRCNN
| maskrcnn
| boxLabelDatastore
| Experiment Manager (Deep Learning Toolbox)