We compare the performance of using deterministic binary neurons (DBNs) and stochastic binary neurons (SBNs) in the proposed model.
proposed model with DBNs | proposed model with SBNs |
---|---|
The following figures show the preactivated outputs (i.e., the real-valued, intermediate values right before the binarization operation).
proposed model with DBNs | proposed model with SBNs |
---|---|
In order to see how DBNs and SBNs work differently, we compute the histograms of their preactivated outputs.
We can see that
We compare the proposed model with a variant that uses normal neurons at the output layer (with sigmoid functions as the activation functions). We refer t this model as the real-valued model.
raw prediction | hard thresholding | Bernoulli sampling |
---|---|---|
We also show the histogram of its probabilistic predictions in the figure above. We can see that
From here we can see how different binarization strategies can shape the characteristics of the preactivated outputs of binary neurons. This also emphasizes the importance of including the binarization operations in the training so that the binarization operations themselves can also be optimized.
We compare the proposed model trained by the WGAN-GP objective with that trained by the GAN objective and by the WGAN objective.
GAN model with DBNs | GAN model with SBNs |
---|---|
WGAN model with DBNs | WGAN model with SBNs |
---|---|
We can see that the WGAN model is able to generate digits with similar qualities as the WGAN-GP model does, while the GAN model suffers from the so-called mode collapse issue.
We compare the performance of using multilayer perceptrons (MLPs) and convolutional neural networks (CNNs). Note that the number of trainable parameters for the MLP and CNN models are 0.53M and 1.4M, respectively.
CNN model with DBNs | CNN model with SBNs |
---|---|
We can see that the CNN model can better capture the characteristics of different digits and generate less artifacts even with a smaller number of trainable parameters as compared to the MLP model.