-
Notifications
You must be signed in to change notification settings - Fork 7
Model files
Gustavo Rosa edited this page Feb 24, 2017
·
9 revisions
This document aims at explaining the model file used for each deep learning technique. Let's get started!
- Restricted Boltzmann Machines (RBM). Suppose we have the following model file:
- Dropout Restricted Boltzmann Machines (Dropout RBM). Suppose we have the following model file:
- Dropconnect Restricted Boltzmann Machines (Dropconnect RBM). Suppose we have the following model file:
- Discriminative Restricted Boltzmann Machines (DRBM). Suppose we have the following model file:
- Gaussian Discriminative Restricted Boltzmann Machines (Gaussian DRBM). Suppose we have the following model file:
- Deep Boltzmann Machines (DBM). Suppose we have the following model file:
- Dropout Deep Boltzmann Machines (Dropout DBM). Suppose we have the following model file:
- Dropconnect Deep Boltzmann Machines (Dropconnect DBM). Suppose we have the following model file:
- Deep Belief Networks (DBN). Suppose we have the following model file:
- Dropout Deep Belief Networks (Dropout DBN). Suppose we have the following model file:
- Dropconnect Deep Belief Networks (Dropconnect DBN). Suppose we have the following model file:
- Enhanced Probabilistic Neural Network (EPNN). Suppose we have the following model file:
100 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
The next line configures RBM minimum and maximum learning rate.
The same model file can be seen for Gaussian RBMs.
100 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
The same model file can be seen for Dropout Gaussian RBMs, Dropout DRBMs and Dropout Gaussian DRBMs.
100 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
0.5 #<dropconnect mask rate>
100 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
The next line configures DRBM minimum and maximum learning rate.
100 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
500 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
500 0.1 0.1 0.00001 #<number of hidden units> <learning rate> <weight decay> <momentum>
0.1 0.9 #<minimum learning rate> <maximum learning rate>
If there is a need in adding more layers, you just need to copy and paste the first two lines from the model. Hence, for example, a DBM with 5 layers will have a model file with 10 lines. Note that this model file can also be used for TDBM (Temperature-based Deep Boltzmann Machines).
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
500 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.5 #<dropconnect mask rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.5 #<dropconnect mask rate>
500 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.8 #<dropconnect mask rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
500 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
500 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
1 #<hidden units dropout rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.5 #<dropconnect mask rate>
100 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.5 #<dropconnect mask rate>
500 0.1 0.0002 0.5 #<number of hidden units> <learning rate> <weight decay> <momentum>
0 0.1 #<minimum learning rate> <maximum learning rate>
0.8 #<dropconnect mask rate>
10 1 0.5 0 # <kmax | 0 for number of labels in training set> <sigma> <radius | greater than 0 for Hyper-Sphere use in EPNN> <learning best parameters | 0 for no-optimization | 1 for grid-search | 2 for grid-search and train/eval sets merging>
Finally, the learning best parameters stands for the desired type of optimization activity. 0 for no-optimization, 1 for a grid-search optimization and 2 for grid-search with training/evaluaiton sets merging.