|
| SoftmaxLayer (CudaDnn< T > cuda, Log log, LayerParameter p) |
| The SoftmaxLayer constructor. More...
|
|
override void | LayerSetUp (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Setup the layer to run in either Engine.CAFFE or Engine.CUDNN mode. More...
|
|
override void | Reshape (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Reshape the bottom (input) and top (output) blobs. More...
|
|
| Layer (CudaDnn< T > cuda, Log log, LayerParameter p) |
| The Layer constructor. More...
|
|
void | Dispose () |
| Releases all GPU and host resources used by the Layer. More...
|
|
virtual void | ConnectLoss (LossLayer< T > layer) |
| Called to connect the loss OnLoss event to a specified layer (typically the data layer). More...
|
|
virtual BlobCollection< T > | PreProcessInput (PropertySet customInput, out int nSeqLen, BlobCollection< T > colBottom=null) |
| The PreprocessInput allows derivative data layers to convert a property set of input data into the bottom blob collection used as intput. More...
|
|
virtual bool | PreProcessInput (string strEncInput, int? nDecInput, BlobCollection< T > colBottom) |
| Preprocess the input data for the RUN phase. More...
|
|
virtual List< Tuple< string, int, double > > | PostProcessOutput (Blob< T > blobSofmtax, int nK=1) |
| The PostProcessOutput allows derivative data layers to post-process the results, converting them back into text results (e.g., detokenizing). More...
|
|
virtual List< Tuple< string, int, double > > | PostProcessLogitsOutput (int nCurIdx, Blob< T > blobLogits, Layer< T > softmax, int nAxis, int nK=1) |
| The PostProcessLogitsOutput allows derivative data layers to post-process the results, converting them back into text results (e.g., detokenizing). More...
|
|
virtual string | PostProcessFullOutput (Blob< T > blobSoftmax) |
| The PostProcessFullOutput allows derivative data layers to post-process the results, usually be detokenizing the data in the blobSoftmax. More...
|
|
virtual string | PostProcessOutput (int nIdx) |
| Convert the index to the word. More...
|
|
virtual void | SetOnDebug (EventHandler< GetWorkBlobArgs< T > > fn) |
| Set the OnDebug event. More...
|
|
virtual void | ResetOnDebug (EventHandler< GetWorkBlobArgs< T > > fn) |
| Reset the OnDebug event, disabling it. More...
|
|
virtual bool | ReInitializeParameters (WEIGHT_TARGET target) |
| Re-initialize the parameters of the layer. More...
|
|
void | SetNetReshapeRequest () |
| Called by the Net when requesting a reshape. More...
|
|
void | SetPhase (Phase phase) |
| Changes the layer's Phase to the one specified. More...
|
|
void | Setup (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Implements common Layer setup functionality. More...
|
|
virtual void | SetNetParameterUsed (NetParameter np) |
| This function allows other layers to gather needed information from the NetParameters if any, and is called when initialzing the Net. More...
|
|
void | ConvertToBase (BlobCollection< T > col) |
| ConvertToBase converts any blobs in a collection that are in half size to the base size. More...
|
|
double | Forward (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Given the bottom (input) Blobs, this function computes the top (output) Blobs and the loss. More...
|
|
void | Backward (BlobCollection< T > colTop, List< bool > rgbPropagateDown, BlobCollection< T > colBottom) |
| Given the top Blob error gradients, compute the bottom Blob error gradients. More...
|
|
double | loss (int nTopIdx) |
| Returns the scalar loss associated with the top Blob at a given index. More...
|
|
void | set_loss (int nTopIdx, double dfLoss) |
| Sets the loss associated with a top Blob at a given index. More...
|
|
virtual bool | AllowForceBackward (int nBottomIdx) |
| Return whether to allow More...
|
|
bool | param_propagate_down (int nParamIdx) |
| Returns whether or not the Layer should compute gradients w.r.t. a parameter at a particular index given by a parameter index. More...
|
|
void | set_param_propagate_down (int nParamIdx, bool bPropagate) |
| Sets whether or not the Layer should compute gradients w.r.t. a parameter at a particular index given by a parameter index. More...
|
|
void | SetEnablePassthrough (bool bEnable) |
| Enables/disables the pass-through mode. More...
|
|
|
override void | dispose () |
| Releases all GPU and host resources used by the Layer. More...
|
|
override void | setup_internal_blobs (BlobCollection< T > col) |
| Derivative layers should add all internal blobws to the 'col' provided. More...
|
|
override void | forward (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Computes the forward calculation using either the Engine.CAFFE or Engine.CUDNN mode. More...
|
|
override void | backward (BlobCollection< T > colTop, List< bool > rgbPropagateDown, BlobCollection< T > colBottom) |
| Computes the error gradient w.r.t the inputs using either the Engine.CAFFE or Engine.CUDNN mode. More...
|
|
void | forward_cuda (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Computes the forward calculation using the Engine.CAFFE mode. More...
|
|
void | backward_cuda (BlobCollection< T > colTop, List< bool > rgbPropagateDown, BlobCollection< T > colBottom) |
| Computes the error gradient w.r.t the inputs using either the Engine.CAFFE. More...
|
|
void | forward_cudnn (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Computes the forward calculation using the Engine.CUDNN mode. More...
|
|
void | backward_cudnn (BlobCollection< T > colTop, List< bool > rgbPropagateDown, BlobCollection< T > colBottom) |
| Computes the error gradient w.r.t the inputs using either the Engine.CUDNN. More...
|
|
void | dispose (ref Layer< T > l) |
| Helper method used to dispose internal layers. More...
|
|
void | dispose (ref Blob< T > b) |
| Helper method used to dispose internal blobs. More...
|
|
void | dispose (ref BlobCollection< T > rg, bool bSetToNull=true) |
| Dispose the blob collection. More...
|
|
GetIterationArgs | getCurrentIteration () |
| Fires the OnGetIteration event to query the current iteration. More...
|
|
long | convert_to_full (int nCount, long hMem) |
| Convert half memory to full memory. More...
|
|
void | convert (BlobCollection< T > col) |
| Convert a collection of blobs from / to half size. More...
|
|
virtual bool | reshapeNeeded (BlobCollection< T > colBottom, BlobCollection< T > colTop, bool bReset=true) |
| Tests the shapes of both the bottom and top blobs and if they are the same as the previous sizing, returns false indicating that no reshape is needed. More...
|
|
bool | compareShapes (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Compare the shapes of the top and bottom and if the same, return true, otherwise false. More...
|
|
void | setShapes (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Set the internal shape sizes - used when determining if a Reshape is necessary. More...
|
|
void | CheckBlobCounts (BlobCollection< T > colBottom, BlobCollection< T > colTop) |
| Called by the Layer::Setup function to check the number of bottom (input) and top (output) Blobs provided match the expected number of blobs expected via the {EactNum,Min,Max}{Bottom,Top}Blobs functions. More...
|
|
void | SetLossWeights (BlobCollection< T > colTop) |
| Called by Layer::Setup to initialize the weights associated with any top (output) Blobs in the loss function ans store non-zero loss weights in the diff Blob. More...
|
|
LayerParameter | convertLayerParam (LayerParameter pChild, LayerParameter pParent) |
| Called to convert a parent LayerParameterEx, used in blob sharing, with a child layer parameter. More...
|
|
bool | shareParameter (Blob< T > b, List< int > rgMinShape, bool bAllowEndsWithComparison=false) |
| Attempts to share a parameter Blob if another parameter Blob with the same name and accpetable size is found. More...
|
|
bool | shareLayerBlob (Blob< T > b, List< int > rgMinShape) |
| Attempts to share a Layer Blob if another parameter Blob with the same name and acceptable size is found. More...
|
|
bool | shareLayerBlobs (Layer< T > layer) |
| Attempts to share the Layer blobs and internal_blobs with matching names and sizes with those in another matching layer. More...
|
|
virtual WorkspaceArgs | getWorkspace () |
| Returns the WorkspaceArgs used to share a workspace between Layers. More...
|
|
virtual bool | setWorkspace (ulong lSizeInBytes) |
| Sets the workspace size (in items) and returns true if set, false otherwise. More...
|
|
void | check_nan (Blob< T > b) |
| Checks a Blob for NaNs and throws an exception if found. More...
|
|
T | convert (double df) |
| Converts a double to a generic. More...
|
|
T | convert (float f) |
| Converts a float to a generic. More...
|
|
double | convertD (T df) |
| Converts a generic to a double value. More...
|
|
float | convertF (T df) |
| Converts a generic to a float value. More...
|
|
double[] | convertD (T[] rg) |
| Converts an array of generic values into an array of double values. More...
|
|
T[] | convert (double[] rg) |
| Converts an array of double values into an array of generic values. More...
|
|
float[] | convertF (T[] rg) |
| Converts an array of float values into an array of generic values. More...
|
|
T[] | convert (float[] rg) |
| Converts an array of float values into an array of generic values. More...
|
|
int | val_at (T[] rg, int nIdx) |
| Returns the integer value at a given index in a generic array. More...
|
|
Size | size_at (Blob< T > b) |
| Returns the Size of a given two element Blob, such as one that stores Blob size information. More...
|
|
The SoftmaxLayer computes the softmax function. This layer is initialized with the MyCaffe.param.SoftmaxParameter.
- See also
- Improving neural networks by preventing co-adaptation of feature detectors by Geoffrey E. Hinton, Nitish Srivastava, Alex Krizhevsky, Ilya Sutskever, and Ruslan R. Salakhutdinov, 2012.
-
Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation by Wu, et al., 2016.
-
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer by Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean, 2017.
-
Exploring the Limits of Language Modeling by Rafal Jozefowicz, Oriol Vinyals, Mike Schuster, Noam Shazeer, and Yonghui Wu, 2016.
- Template Parameters
-
T | Specifies the base type float or double. Using float is recommended to conserve GPU memory. |
Definition at line 23 of file SoftmaxLayer.cs.