From 89ee010a94ebe4ce84433ac0f927e007978df9be Mon Sep 17 00:00:00 2001 From: Dario Coscia <93731561+dario-coscia@users.noreply.github.com> Date: Fri, 7 Jun 2024 18:07:24 +0200 Subject: [PATCH] Update feed_forward.py (#305) --- pina/model/feed_forward.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pina/model/feed_forward.py b/pina/model/feed_forward.py index ba98680..5dfd791 100644 --- a/pina/model/feed_forward.py +++ b/pina/model/feed_forward.py @@ -20,7 +20,7 @@ class FeedForward(torch.nn.Module): :param int inner_size: number of neurons in the hidden layer(s). Default is 20. :param int n_layers: number of hidden layers. Default is 2. - :param func: the activation function to use. If a single + :param torch.nn.Module func: the activation function to use. If a single :class:`torch.nn.Module` is passed, this is used as activation function after any layers, except the last one. If a list of Modules is passed, they are used as activation functions at any layers, in order. @@ -116,7 +116,7 @@ class ResidualFeedForward(torch.nn.Module): :param int inner_size: number of neurons in the hidden layer(s). Default is 20. :param int n_layers: number of hidden layers. Default is 2. - :param func: the activation function to use. If a single + :param torch.nn.Module func: the activation function to use. If a single :class:`torch.nn.Module` is passed, this is used as activation function after any layers, except the last one. If a list of Modules is passed, they are used as activation functions at any layers, in order.