Codacy Small Bug Fixes:
- cleaned up imports - cleaned up some code - added docstrings
This commit is contained in:
@@ -1,3 +1,4 @@
|
||||
""" Implementation of adaptive linear layer. """
|
||||
import torch
|
||||
from torch.nn.parameter import Parameter
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
import torch
|
||||
from torch.nn.parameter import Parameter
|
||||
|
||||
class AdaptiveReLU(torch.nn.Module):
|
||||
class AdaptiveReLU(torch.nn.Module, Parameter):
|
||||
'''
|
||||
Implementation of soft exponential activation.
|
||||
Shape:
|
||||
|
||||
Reference in New Issue
Block a user