Welcome to mirror list, hosted at ThFree Co, Russian Federation.

SoftPlus.lua - github.com/torch/nn.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
blob: f77b25380002502a896d7d59255a68a470e00246 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
local SoftPlus, parent = torch.class('nn.SoftPlus', 'nn.Module')

function SoftPlus:__init(beta)
   parent.__init(self)
   self.beta = beta or 1  -- Beta controls sharpness of transfer function
   self.threshold = 20    -- Avoid floating point issues with exp(x), x>20
end

function SoftPlus:updateOutput(input)
   -- f(x) = 1/beta * log(1 + exp(beta * x))
   input.THNN.SoftPlus_updateOutput(
      input:cdata(),
      self.output:cdata(),
      self.beta,
      self.threshold
   )
   return self.output
end

function SoftPlus:updateGradInput(input, gradOutput)
   -- d/dx[log(1+exp(k*x))/k] = exp(kx) / (exp(kx) + 1)
   -- SINCE
   -- y = (1/k)*log(1+exp(k*x)) --> x = (1/k)*log(exp(k*y)-1)
   -- THEREFORE:
   -- d/dx(f(x)) = (exp(k*y) - 1) / exp(k*y)
   input.THNN.SoftPlus_updateGradInput(
      input:cdata(),
      gradOutput:cdata(),
      self.gradInput:cdata(),
      self.output:cdata(),
      self.beta,
      self.threshold
   )
   return self.gradInput
end