[docs]classSiLU(nn.Module):"""Implements the SiLU activation from `"Gaussian Error Linear Units (GELUs)" <https://arxiv.org/pdf/1606.08415.pdf>`_ (also known as Swish). This activation is computed as follows: .. math:: f(x) = x \\cdot \\sigma(x) """defforward(self,x:Tensor)->Tensor:return_SiLU.apply(x)
[docs]classMish(nn.Module):"""Implements the Mish activation module from `"Mish: A Self Regularized Non-Monotonic Neural Activation Function" <https://arxiv.org/pdf/1908.08681.pdf>`_ This activation is computed as follows: .. math:: f(x) = x \\cdot \\tanh(ln(1 + e^x)) """defforward(self,x:Tensor)->Tensor:returnF.mish(x)
[docs]classHardMish(_Activation):"""Implements the Had Mish activation module from `"H-Mish" <https://github.com/digantamisra98/H-Mish>`_ This activation is computed as follows: .. math:: f(x) = \\frac{x}{2} \\cdot \\min(2, \\max(0, x + 2)) """defforward(self,x:Tensor)->Tensor:returnF.hard_mish(x,inplace=self.inplace)
[docs]classNLReLU(_Activation):"""Implements the Natural-Logarithm ReLU activation module from `"Natural-Logarithm-Rectified Activation Function in Convolutional Neural Networks" <https://arxiv.org/pdf/1908.03682.pdf>`_ This activation is computed as follows: .. math:: f(x) = ln(1 + \\beta \\cdot max(0, x)) Args: inplace (bool): should the operation be performed inplace """defforward(self,x:Tensor)->Tensor:returnF.nl_relu(x,inplace=self.inplace)
[docs]classFReLU(nn.Module):"""Implements the Funnel activation module from `"Funnel Activation for Visual Recognition" <https://arxiv.org/pdf/2007.11824.pdf>`_ This activation is computed as follows: .. math:: f(x) = max(\\mathbb{T}(x), x) where the :math:`\\mathbb{T}` is the spatial contextual feature extraction. It is a convolution filter of size `kernel_size`, same padding and groups equal to the number of input channels, followed by a batch normalization. Args: inplace (bool): should the operation be performed inplace """def__init__(self,in_channels:int,kernel_size:int=3)->None:super().__init__()self.conv=nn.Conv2d(in_channels,in_channels,kernel_size,padding=kernel_size//2,groups=in_channels)self.bn=nn.BatchNorm2d(in_channels)defforward(self,x:Tensor)->Tensor:out=self.conv(x)out=self.bn(out)x=torch.max(x,out)returnx