gated_sum
allennlp.modules.gated_sum
GatedSum#
class GatedSum(torch.nn.Module):
| def __init__(
| self,
| input_dim: int,
| activation: Activation = torch.nn.Sigmoid()
| ) -> None
This Module
represents a gated sum of two tensors a
and b
. Specifically:
f = activation(W [a; b])
out = f * a + (1 - f) * b
Parameters
- input_dim :
int
The dimensionality of the input. We assume the input have shape(..., input_dim)
. - activation :
Activation
, optional (default =torch.nn.Sigmoid()
)
The activation function to use.
get_input_dim#
class GatedSum(torch.nn.Module):
| ...
| def get_input_dim(self)
get_output_dim#
class GatedSum(torch.nn.Module):
| ...
| def get_output_dim(self)
forward#
class GatedSum(torch.nn.Module):
| ...
| def forward(
| self,
| input_a: torch.Tensor,
| input_b: torch.Tensor
| ) -> torch.Tensor