Table 1:

Architecture . | Representative Models . | Model Features . |
---|---|---|

Deep belief net | RBM (Zhang et al., 2018) | A generative graphic model that uses the energy to capture the probability distribution between visible units and hidden units. |

SRBM (Chen et al., 2017) | A sparse variant that each hidden unit connects to part of the visible units, preventing the model overfitting based on hierarchical latent tree analysis. | |

FRBM (Ning et al., 2018) | A fast variant trained by the lean CD algorithm in which the bounds-based filtering and delta product reduce the redundant dot product calculations. | |

TTRBM (Ju et al., 2019) | A compact variant that the parameters between the visible layer and hidden layer are reduced by transforming into the tensor-train format. | |

Stacked autoencoder | AE (Michael et al., 2018) | A basic fully connected network that uses the encoder-decoder strategy in an unsupervised manner to learn intrinsic features of data. |

DAE (Vincent et al., 2008) | A denoising variant that reconstructs the clear data from the noising data. | |

SAE (Makhzani & Frey, 2013) | A sparse variant that captures the sparse representations of the input by adding the constraint into the loss function. | |

GAE (Hou et al., 2019) | An adversarial variant that the decoder subnetwork that is also regarded as the generator, adopting game theory to more consistent features with input data. | |

FAE (Ashfahani et al., 2019) | An evolving variant that constructs an adaptive network structure in the learning of representations, based on the network significance. | |

BAE (Angshul, 2019) | An evolving variant adding the path-loss term in the loss function based on dictionary learning. | |

Convolutional neural network | Alexnet (Krizhevsky, Sutskever, & Hinton, 2012) | The nonsaturating neurons and the dropout are adopted in the nonlinear computational layers, based on a GPU implementation, respectively. |

ResNet (He et al., 2016) | A shortcut connection is used to cross several layers to back propagate the network loss to previous layers. | |

Inception (Christian et al., 2017) | A deeper and wider network is designed by using the uniform grid size for the blocks with auxiliary information. | |

SEnet (Cao et al., 2019) | Informational embedding and adaption recalibration are regarded as self-attention operations. | |

ECNN (Sandler et al., 2018) | The low-rank convolution replaces the full-rank convolution to improve the learning efficiency without much accuracy loss. | |

Recurrent neural network | RNN (Zhang et al., 2014) | A fully connected network where the self-connection between hidden layers is used to model the time dependency. |

BiRNN (Schuster & Paliwal, 1997) | Two independent computing processes are used to encode the forward and the backward dependency. | |

LSTM (Hochreiter & Schmidhuber, 1997) | The memory block is introduced to model the long-time dependency well. | |

SRNN (Lei et al., 2018) | A fast variant in which the light recurrence and highway network are proposed to improve the learning efficiency for a parallelized implementation. | |

VRNN (Jang et al., 2019) | A variational variant that uses the variational encoder-decoder strategy to model the temporal intrinsic features. |

Architecture . | Representative Models . | Model Features . |
---|---|---|

Deep belief net | RBM (Zhang et al., 2018) | A generative graphic model that uses the energy to capture the probability distribution between visible units and hidden units. |

SRBM (Chen et al., 2017) | A sparse variant that each hidden unit connects to part of the visible units, preventing the model overfitting based on hierarchical latent tree analysis. | |

FRBM (Ning et al., 2018) | A fast variant trained by the lean CD algorithm in which the bounds-based filtering and delta product reduce the redundant dot product calculations. | |

TTRBM (Ju et al., 2019) | A compact variant that the parameters between the visible layer and hidden layer are reduced by transforming into the tensor-train format. | |

Stacked autoencoder | AE (Michael et al., 2018) | A basic fully connected network that uses the encoder-decoder strategy in an unsupervised manner to learn intrinsic features of data. |

DAE (Vincent et al., 2008) | A denoising variant that reconstructs the clear data from the noising data. | |

SAE (Makhzani & Frey, 2013) | A sparse variant that captures the sparse representations of the input by adding the constraint into the loss function. | |

GAE (Hou et al., 2019) | An adversarial variant that the decoder subnetwork that is also regarded as the generator, adopting game theory to more consistent features with input data. | |

FAE (Ashfahani et al., 2019) | An evolving variant that constructs an adaptive network structure in the learning of representations, based on the network significance. | |

BAE (Angshul, 2019) | An evolving variant adding the path-loss term in the loss function based on dictionary learning. | |

Convolutional neural network | Alexnet (Krizhevsky, Sutskever, & Hinton, 2012) | The nonsaturating neurons and the dropout are adopted in the nonlinear computational layers, based on a GPU implementation, respectively. |

ResNet (He et al., 2016) | A shortcut connection is used to cross several layers to back propagate the network loss to previous layers. | |

Inception (Christian et al., 2017) | A deeper and wider network is designed by using the uniform grid size for the blocks with auxiliary information. | |

SEnet (Cao et al., 2019) | Informational embedding and adaption recalibration are regarded as self-attention operations. | |

ECNN (Sandler et al., 2018) | The low-rank convolution replaces the full-rank convolution to improve the learning efficiency without much accuracy loss. | |

Recurrent neural network | RNN (Zhang et al., 2014) | A fully connected network where the self-connection between hidden layers is used to model the time dependency. |

BiRNN (Schuster & Paliwal, 1997) | Two independent computing processes are used to encode the forward and the backward dependency. | |

LSTM (Hochreiter & Schmidhuber, 1997) | The memory block is introduced to model the long-time dependency well. | |

SRNN (Lei et al., 2018) | A fast variant in which the light recurrence and highway network are proposed to improve the learning efficiency for a parallelized implementation. | |

VRNN (Jang et al., 2019) | A variational variant that uses the variational encoder-decoder strategy to model the temporal intrinsic features. |

Notes: RBM: restricted Boltzmann machine; SRBM: sparse restricted Boltzmann machine; FRBM: fast restricted Boltzmann machine; TTRBM: tensor-train restricted Boltzmann machine; AE: autoencoder; DAE: denoising autoencoder; SAE: K-sparse autoencoder; GAE: generative autoencoder; FAE: fast autoencoder; BAE: blind autoencoder; Alexnet: Alex convolutional net; ResNet: residual convolutional net; Inception: Inception; SEnet: squeeze excitation network; ECNN: efficient convolutional neural network; RNN: recurrent neural network; BiRNN: bidirectional recurrent neural network; LSTM: long short-term memory; SRNN: slight recurrent neural network; VRNN: variational recurrent neural network.

This site uses cookies. By continuing to use our website, you are agreeing to our privacy policy.