Open
Description
Hi, I noticed that the activations are not binary, but floats between 0 and 1, and I was wandering if there is a bug.
The usage of floats is due to the fact that, also in the binary models, the hard tanh function is used, e.g.:
self.tanh2 = nn.Hardtanh(inplace=True)
In the paper, however, it is mentioned that the activation function should behave as a sign function in the forward step - is this correct? Thanks,
Metadata
Metadata
Assignees
Labels
No labels