Improve differential operators (#528)
* Improve grad logic and fix issues * Add operators' fast versions * Fix bug in laplacian + new tests + restructuring Co-authored-by: Dario Coscia <dariocos99@gmail.com> * fix advection bug --------- Co-authored-by: Dario Coscia <dariocos99@gmail.com>
This commit is contained in:
committed by
FilippoOlivo
parent
938bdeb421
commit
fa6fda0bd5
564
pina/operator.py
564
pina/operator.py
@@ -10,10 +10,309 @@ Each differential operator takes the following inputs:
|
||||
- A tensor with respect to which the operator is computed.
|
||||
- The names of the output variables for which the operator is evaluated.
|
||||
- The names of the variables with respect to which the operator is computed.
|
||||
|
||||
Each differential operator has its fast version, which performs no internal
|
||||
checks on input and output tensors. For these methods, the user is always
|
||||
required to specify both ``components`` and ``d`` as lists of strings.
|
||||
"""
|
||||
|
||||
import torch
|
||||
from pina.label_tensor import LabelTensor
|
||||
from .label_tensor import LabelTensor
|
||||
|
||||
|
||||
def _check_values(output_, input_, components, d):
|
||||
"""
|
||||
Perform checks on arguments of differential operators.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the operator is
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
operator is computed.
|
||||
:param components: The names of the output variables for which to compute
|
||||
the operator. It must be a subset of the output labels.
|
||||
If ``None``, all output variables are considered. Default is ``None``.
|
||||
:type components: str | list[str]
|
||||
:param d: The names of the input variables with respect to which the
|
||||
operator is computed. It must be a subset of the input labels.
|
||||
If ``None``, all input variables are considered. Default is ``None``.
|
||||
:type d: str | list[str]
|
||||
:raises TypeError: If the input tensor is not a LabelTensor.
|
||||
:raises TypeError: If the output tensor is not a LabelTensor.
|
||||
:raises RuntimeError: If derivative labels are missing from the ``input_``.
|
||||
:raises RuntimeError: If component labels are missing from the ``output_``.
|
||||
:return: The components and d lists.
|
||||
:rtype: tuple[list[str], list[str]]
|
||||
"""
|
||||
# Check if the input is a LabelTensor
|
||||
if not isinstance(input_, LabelTensor):
|
||||
raise TypeError("Input must be a LabelTensor.")
|
||||
|
||||
# Check if the output is a LabelTensor
|
||||
if not isinstance(output_, LabelTensor):
|
||||
raise TypeError("Output must be a LabelTensor.")
|
||||
|
||||
# If no labels are provided, use all labels
|
||||
d = d or input_.labels
|
||||
components = components or output_.labels
|
||||
|
||||
# Convert to list if not already
|
||||
d = d if isinstance(d, list) else [d]
|
||||
components = components if isinstance(components, list) else [components]
|
||||
|
||||
# Check if all labels are present in the input tensor
|
||||
if not all(di in input_.labels for di in d):
|
||||
raise RuntimeError("Derivative labels missing from input tensor.")
|
||||
|
||||
# Check if all labels are present in the output tensor
|
||||
if not all(c in output_.labels for c in components):
|
||||
raise RuntimeError("Component label missing from output tensor.")
|
||||
|
||||
return components, d
|
||||
|
||||
|
||||
def _scalar_grad(output_, input_, d):
|
||||
"""
|
||||
Compute the gradient of a scalar-valued ``output_``.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the gradient is
|
||||
computed. It must be a column tensor.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
gradient is computed.
|
||||
:param list[str] d: The names of the input variables with respect to
|
||||
which the gradient is computed. It must be a subset of the input
|
||||
labels. If ``None``, all input variables are considered.
|
||||
:return: The computed gradient tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
grad_out = torch.autograd.grad(
|
||||
outputs=output_,
|
||||
inputs=input_,
|
||||
grad_outputs=torch.ones_like(output_),
|
||||
create_graph=True,
|
||||
retain_graph=True,
|
||||
allow_unused=True,
|
||||
)[0]
|
||||
|
||||
return grad_out[..., [input_.labels.index(i) for i in d]]
|
||||
|
||||
|
||||
def _scalar_laplacian(output_, input_, d):
|
||||
"""
|
||||
Compute the laplacian of a scalar-valued ``output_``.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the laplacian is
|
||||
computed. It must be a column tensor.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
laplacian is computed.
|
||||
:param list[str] d: The names of the input variables with respect to
|
||||
which the laplacian is computed. It must be a subset of the input
|
||||
labels. If ``None``, all input variables are considered.
|
||||
:return: The computed laplacian tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
first_grad = fast_grad(
|
||||
output_=output_, input_=input_, components=output_.labels, d=d
|
||||
)
|
||||
second_grad = fast_grad(
|
||||
output_=first_grad, input_=input_, components=first_grad.labels, d=d
|
||||
)
|
||||
labels_to_extract = [f"d{c}d{d_}" for c, d_ in zip(first_grad.labels, d)]
|
||||
return torch.sum(
|
||||
second_grad.extract(labels_to_extract), dim=-1, keepdim=True
|
||||
)
|
||||
|
||||
|
||||
def fast_grad(output_, input_, components, d):
|
||||
"""
|
||||
Compute the gradient of the ``output_`` with respect to the ``input``.
|
||||
|
||||
Unlike ``grad``, this function performs no internal checks on input and
|
||||
output tensors. The user is required to specify both ``components`` and
|
||||
``d`` as lists of strings. It is designed to enhance computation speed.
|
||||
|
||||
This operator supports both vector-valued and scalar-valued functions with
|
||||
one or multiple input coordinates.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the gradient is
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
gradient is computed.
|
||||
:param list[str] components: The names of the output variables for which to
|
||||
compute the gradient. It must be a subset of the output labels.
|
||||
:param list[str] d: The names of the input variables with respect to which
|
||||
the gradient is computed. It must be a subset of the input labels.
|
||||
:return: The computed gradient tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
# Scalar gradient
|
||||
if output_.shape[-1] == 1:
|
||||
return LabelTensor(
|
||||
_scalar_grad(output_=output_, input_=input_, d=d),
|
||||
labels=[f"d{output_.labels[0]}d{i}" for i in d],
|
||||
)
|
||||
|
||||
# Vector gradient
|
||||
grads = torch.cat(
|
||||
[
|
||||
_scalar_grad(output_=output_.extract(c), input_=input_, d=d)
|
||||
for c in components
|
||||
],
|
||||
dim=-1,
|
||||
)
|
||||
|
||||
return LabelTensor(
|
||||
grads, labels=[f"d{c}d{i}" for c in components for i in d]
|
||||
)
|
||||
|
||||
|
||||
def fast_div(output_, input_, components, d):
|
||||
"""
|
||||
Compute the divergence of the ``output_`` with respect to ``input``.
|
||||
|
||||
Unlike ``div``, this function performs no internal checks on input and
|
||||
output tensors. The user is required to specify both ``components`` and
|
||||
``d`` as lists of strings. It is designed to enhance computation speed.
|
||||
|
||||
This operator supports vector-valued functions with multiple input
|
||||
coordinates.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the divergence is
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
divergence is computed.
|
||||
:param list[str] components: The names of the output variables for which to
|
||||
compute the divergence. It must be a subset of the output labels.
|
||||
:param list[str] d: The names of the input variables with respect to which
|
||||
the divergence is computed. It must be a subset of the input labels.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
grad_out = fast_grad(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
tensors_to_sum = [
|
||||
grad_out.extract(f"d{c}d{d_}") for c, d_ in zip(components, d)
|
||||
]
|
||||
|
||||
return LabelTensor.summation(tensors_to_sum)
|
||||
|
||||
|
||||
def fast_laplacian(output_, input_, components, d, method="std"):
|
||||
"""
|
||||
Compute the laplacian of the ``output_`` with respect to ``input``.
|
||||
|
||||
Unlike ``laplacian``, this function performs no internal checks on input and
|
||||
output tensors. The user is required to specify both ``components`` and
|
||||
``d`` as lists of strings. It is designed to enhance computation speed.
|
||||
|
||||
This operator supports both vector-valued and scalar-valued functions with
|
||||
one or multiple input coordinates.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the laplacian is
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
laplacian is computed.
|
||||
:param list[str] components: The names of the output variables for which to
|
||||
compute the laplacian. It must be a subset of the output labels.
|
||||
:param list[str] d: The names of the input variables with respect to which
|
||||
the laplacian is computed. It must be a subset of the input labels.
|
||||
:param str method: The method used to compute the Laplacian. Available
|
||||
methods are ``std`` and ``divgrad``. The ``std`` method computes the
|
||||
trace of the Hessian matrix, while the ``divgrad`` method computes the
|
||||
divergence of the gradient. Default is ``std``.
|
||||
:return: The computed laplacian tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
# Scalar laplacian
|
||||
if output_.shape[-1] == 1:
|
||||
return LabelTensor(
|
||||
_scalar_laplacian(output_=output_, input_=input_, d=d),
|
||||
labels=[f"dd{c}" for c in components],
|
||||
)
|
||||
|
||||
# Initialize the result tensor and its labels
|
||||
labels = [f"dd{c}" for c in components]
|
||||
result = torch.empty(
|
||||
input_.shape[0], len(components), device=output_.device
|
||||
)
|
||||
|
||||
# Vector laplacian
|
||||
if method == "std":
|
||||
result = torch.cat(
|
||||
[
|
||||
_scalar_laplacian(
|
||||
output_=output_.extract(c), input_=input_, d=d
|
||||
)
|
||||
for c in components
|
||||
],
|
||||
dim=-1,
|
||||
)
|
||||
|
||||
elif method == "divgrad":
|
||||
grads = fast_grad(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
result = torch.cat(
|
||||
[
|
||||
fast_div(
|
||||
output_=grads,
|
||||
input_=input_,
|
||||
components=[f"d{c}d{i}" for i in d],
|
||||
d=d,
|
||||
)
|
||||
for c in components
|
||||
],
|
||||
dim=-1,
|
||||
)
|
||||
|
||||
else:
|
||||
raise ValueError(
|
||||
"Invalid method. Available methods are ``std`` and ``divgrad``."
|
||||
)
|
||||
|
||||
return LabelTensor(result, labels=labels)
|
||||
|
||||
|
||||
def fast_advection(output_, input_, velocity_field, components, d):
|
||||
"""
|
||||
Perform the advection operation on the ``output_`` with respect to the
|
||||
``input``. This operator support vector-valued functions with multiple input
|
||||
coordinates.
|
||||
|
||||
Unlike ``advection``, this function performs no internal checks on input and
|
||||
output tensors. The user is required to specify both ``components`` and
|
||||
``d`` as lists of strings. It is designed to enhance computation speed.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the advection is
|
||||
computed.
|
||||
:param LabelTensor input_: the input tensor with respect to which advection
|
||||
is computed.
|
||||
:param str velocity_field: The name of the output variable used as velocity
|
||||
field. It must be chosen among the output labels.
|
||||
:param list[str] components: The names of the output variables for which to
|
||||
compute the advection. It must be a subset of the output labels.
|
||||
:param list[str] d: The names of the input variables with respect to which
|
||||
the advection is computed. It must be a subset of the input labels.
|
||||
:return: The computed advection tensor.
|
||||
:rtype: torch.Tensor
|
||||
"""
|
||||
# Add a dimension to the velocity field for following operations
|
||||
velocity = output_.extract(velocity_field).unsqueeze(-1)
|
||||
|
||||
# Remove the velocity field from the components
|
||||
filter_components = [c for c in components if c != velocity_field]
|
||||
|
||||
# Compute the gradient
|
||||
grads = fast_grad(
|
||||
output_=output_, input_=input_, components=filter_components, d=d
|
||||
)
|
||||
|
||||
# Reshape into [..., len(filter_components), len(d)]
|
||||
tmp = grads.reshape(*output_.shape[:-1], len(filter_components), len(d))
|
||||
|
||||
# Transpose to [..., len(d), len(filter_components)]
|
||||
tmp = tmp.transpose(-1, -2)
|
||||
|
||||
return (tmp * velocity).sum(dim=tmp.tensor.ndim - 2)
|
||||
|
||||
|
||||
def grad(output_, input_, components=None, d=None):
|
||||
@@ -27,95 +326,25 @@ def grad(output_, input_, components=None, d=None):
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
gradient is computed.
|
||||
:param components: The names of the output variables for which to
|
||||
compute the gradient. It must be a subset of the output labels.
|
||||
:param components: The names of the output variables for which to compute
|
||||
the gradient. It must be a subset of the output labels.
|
||||
If ``None``, all output variables are considered. Default is ``None``.
|
||||
:type components: str | list[str]
|
||||
:param d: The names of the input variables with respect to which
|
||||
the gradient is computed. It must be a subset of the input labels.
|
||||
:param d: The names of the input variables with respect to which the
|
||||
gradient is computed. It must be a subset of the input labels.
|
||||
If ``None``, all input variables are considered. Default is ``None``.
|
||||
:type d: str | list[str]
|
||||
:raises TypeError: If the input tensor is not a LabelTensor.
|
||||
:raises RuntimeError: If the output is a scalar field and the components
|
||||
are not equal to the output labels.
|
||||
:raises NotImplementedError: If the output is neither a vector field nor a
|
||||
scalar field.
|
||||
:raises TypeError: If the output tensor is not a LabelTensor.
|
||||
:raises RuntimeError: If derivative labels are missing from the ``input_``.
|
||||
:raises RuntimeError: If component labels are missing from the ``output_``.
|
||||
:return: The computed gradient tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
|
||||
def grad_scalar_output(output_, input_, d):
|
||||
"""
|
||||
Compute the gradient of a scalar-valued ``output_``.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the gradient is
|
||||
computed. It must be a column tensor.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
gradient is computed.
|
||||
:param d: The names of the input variables with respect to
|
||||
which the gradient is computed. It must be a subset of the input
|
||||
labels. If ``None``, all input variables are considered.
|
||||
:type d: str | list[str]
|
||||
:raises RuntimeError: If a vectorial function is passed.
|
||||
:raises RuntimeError: If missing derivative labels.
|
||||
:return: The computed gradient tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
|
||||
if len(output_.labels) != 1:
|
||||
raise RuntimeError("only scalar function can be differentiated")
|
||||
if not all(di in input_.labels for di in d):
|
||||
raise RuntimeError("derivative labels missing from input tensor")
|
||||
|
||||
output_fieldname = output_.labels[0]
|
||||
gradients = torch.autograd.grad(
|
||||
output_,
|
||||
input_,
|
||||
grad_outputs=torch.ones(
|
||||
output_.size(), dtype=output_.dtype, device=output_.device
|
||||
),
|
||||
create_graph=True,
|
||||
retain_graph=True,
|
||||
allow_unused=True,
|
||||
)[0]
|
||||
gradients.labels = input_.stored_labels
|
||||
gradients = gradients[..., [input_.labels.index(i) for i in d]]
|
||||
gradients.labels = [f"d{output_fieldname}d{i}" for i in d]
|
||||
return gradients
|
||||
|
||||
if not isinstance(input_, LabelTensor):
|
||||
raise TypeError
|
||||
|
||||
if d is None:
|
||||
d = input_.labels
|
||||
|
||||
if components is None:
|
||||
components = output_.labels
|
||||
|
||||
if not isinstance(components, list):
|
||||
components = [components]
|
||||
|
||||
if not isinstance(d, list):
|
||||
d = [d]
|
||||
|
||||
if output_.shape[1] == 1: # scalar output ################################
|
||||
|
||||
if components != output_.labels:
|
||||
raise RuntimeError
|
||||
gradients = grad_scalar_output(output_, input_, d)
|
||||
|
||||
elif (
|
||||
output_.shape[output_.ndim - 1] >= 2
|
||||
): # vector output ##############################
|
||||
tensor_to_cat = []
|
||||
for i, c in enumerate(components):
|
||||
c_output = output_.extract([c])
|
||||
tensor_to_cat.append(grad_scalar_output(c_output, input_, d))
|
||||
gradients = LabelTensor.cat(tensor_to_cat, dim=output_.tensor.ndim - 1)
|
||||
else:
|
||||
raise NotImplementedError
|
||||
|
||||
return gradients
|
||||
components, d = _check_values(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
return fast_grad(output_=output_, input_=input_, components=components, d=d)
|
||||
|
||||
|
||||
def div(output_, input_, components=None, d=None):
|
||||
@@ -129,51 +358,31 @@ def div(output_, input_, components=None, d=None):
|
||||
computed.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
divergence is computed.
|
||||
:param components: The names of the output variables for which to
|
||||
compute the divergence. It must be a subset of the output labels.
|
||||
:param components: The names of the output variables for which to compute
|
||||
the divergence. It must be a subset of the output labels.
|
||||
If ``None``, all output variables are considered. Default is ``None``.
|
||||
:type components: str | list[str]
|
||||
:param d: The names of the input variables with respect to which
|
||||
the divergence is computed. It must be a subset of the input labels.
|
||||
:param d: The names of the input variables with respect to which the
|
||||
divergence is computed. It must be a subset of the input labels.
|
||||
If ``None``, all input variables are considered. Default is ``None``.
|
||||
:type d: str | list[str]
|
||||
:type components: str | list[str]
|
||||
:raises TypeError: If the input tensor is not a LabelTensor.
|
||||
:raises ValueError: If the output is a scalar field.
|
||||
:raises ValueError: If the number of components is not equal to the number
|
||||
of input variables.
|
||||
:raises TypeError: If the output tensor is not a LabelTensor.
|
||||
:raises ValueError: If the length of ``components`` and ``d`` do not match.
|
||||
:return: The computed divergence tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
if not isinstance(input_, LabelTensor):
|
||||
raise TypeError
|
||||
|
||||
if d is None:
|
||||
d = input_.labels
|
||||
|
||||
if components is None:
|
||||
components = output_.labels
|
||||
|
||||
if not isinstance(components, list):
|
||||
components = [components]
|
||||
|
||||
if not isinstance(d, list):
|
||||
d = [d]
|
||||
|
||||
if output_.shape[1] < 2 or len(components) < 2:
|
||||
raise ValueError("div supported only for vector fields")
|
||||
components, d = _check_values(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
|
||||
# Components and d must be of the same length
|
||||
if len(components) != len(d):
|
||||
raise ValueError
|
||||
raise ValueError(
|
||||
"Divergence requires components and d to be of the same length."
|
||||
)
|
||||
|
||||
grad_output = grad(output_, input_, components, d)
|
||||
labels = [None] * len(components)
|
||||
tensors_to_sum = []
|
||||
for i, (c, d_) in enumerate(zip(components, d)):
|
||||
c_fields = f"d{c}d{d_}"
|
||||
tensors_to_sum.append(grad_output.extract(c_fields))
|
||||
labels[i] = c_fields
|
||||
div_result = LabelTensor.summation(tensors_to_sum)
|
||||
return div_result
|
||||
return fast_div(output_=output_, input_=input_, components=components, d=d)
|
||||
|
||||
|
||||
def laplacian(output_, input_, components=None, d=None, method="std"):
|
||||
@@ -195,71 +404,22 @@ def laplacian(output_, input_, components=None, d=None, method="std"):
|
||||
the laplacian is computed. It must be a subset of the input labels.
|
||||
If ``None``, all input variables are considered. Default is ``None``.
|
||||
:type d: str | list[str]
|
||||
:param str method: The method used to compute the Laplacian. Default is
|
||||
``std``.
|
||||
:raises NotImplementedError: If ``std=divgrad``.
|
||||
:param str method: The method used to compute the Laplacian. Available
|
||||
methods are ``std`` and ``divgrad``. The ``std`` method computes the
|
||||
trace of the Hessian matrix, while the ``divgrad`` method computes the
|
||||
divergence of the gradient. Default is ``std``.
|
||||
:raises TypeError: If the input tensor is not a LabelTensor.
|
||||
:raises TypeError: If the output tensor is not a LabelTensor.
|
||||
:raises ValueError: If the passed method is neither ``std`` nor ``divgrad``.
|
||||
:return: The computed laplacian tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
|
||||
def scalar_laplace(output_, input_, components, d):
|
||||
"""
|
||||
Compute the laplacian of a scalar-valued ``output_``.
|
||||
|
||||
:param LabelTensor output_: The output tensor on which the laplacian is
|
||||
computed. It must be a column tensor.
|
||||
:param LabelTensor input_: The input tensor with respect to which the
|
||||
laplacian is computed.
|
||||
:param components: The names of the output variables for which
|
||||
to compute the laplacian. It must be a subset of the output labels.
|
||||
If ``None``, all output variables are considered.
|
||||
:type components: str | list[str]
|
||||
:param d: The names of the input variables with respect to
|
||||
which the laplacian is computed. It must be a subset of the input
|
||||
labels. If ``None``, all input variables are considered.
|
||||
:type d: str | list[str]
|
||||
:return: The computed laplacian tensor.
|
||||
:rtype: LabelTensor
|
||||
"""
|
||||
|
||||
grad_output = grad(output_, input_, components=components, d=d)
|
||||
result = torch.zeros(output_.shape[0], 1, device=output_.device)
|
||||
|
||||
for i, label in enumerate(grad_output.labels):
|
||||
gg = grad(grad_output, input_, d=d, components=[label])
|
||||
result[:, 0] += super(torch.Tensor, gg.T).__getitem__(i)
|
||||
|
||||
return result
|
||||
|
||||
if d is None:
|
||||
d = input_.labels
|
||||
|
||||
if components is None:
|
||||
components = output_.labels
|
||||
|
||||
if not isinstance(components, list):
|
||||
components = [components]
|
||||
|
||||
if not isinstance(d, list):
|
||||
d = [d]
|
||||
|
||||
if method == "divgrad":
|
||||
raise NotImplementedError("divgrad not implemented as method")
|
||||
|
||||
if method == "std":
|
||||
|
||||
result = torch.empty(
|
||||
input_.shape[0], len(components), device=output_.device
|
||||
)
|
||||
labels = [None] * len(components)
|
||||
for idx, c in enumerate(components):
|
||||
result[:, idx] = scalar_laplace(output_, input_, [c], d).flatten()
|
||||
labels[idx] = f"dd{c}"
|
||||
|
||||
result = result.as_subclass(LabelTensor)
|
||||
result.labels = labels
|
||||
|
||||
return result
|
||||
components, d = _check_values(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
return fast_laplacian(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
|
||||
|
||||
def advection(output_, input_, velocity_field, components=None, d=None):
|
||||
@@ -274,34 +434,34 @@ def advection(output_, input_, velocity_field, components=None, d=None):
|
||||
is computed.
|
||||
:param str velocity_field: The name of the output variable used as velocity
|
||||
field. It must be chosen among the output labels.
|
||||
:param components: The names of the output variables for which
|
||||
to compute the advection. It must be a subset of the output labels.
|
||||
:param components: The names of the output variables for which to compute
|
||||
the advection. It must be a subset of the output labels.
|
||||
If ``None``, all output variables are considered. Default is ``None``.
|
||||
:type components: str | list[str]
|
||||
:param d: The names of the input variables with respect to which
|
||||
the advection is computed. It must be a subset of the input labels.
|
||||
:param d: The names of the input variables with respect to which the
|
||||
advection is computed. It must be a subset of the input labels.
|
||||
If ``None``, all input variables are considered. Default is ``None``.
|
||||
:type d: str | list[str]
|
||||
:raises TypeError: If the input tensor is not a LabelTensor.
|
||||
:raises TypeError: If the output tensor is not a LabelTensor.
|
||||
:raises RuntimeError: If the velocity field is not in the output labels.
|
||||
:return: The computed advection tensor.
|
||||
:rtype: LabelTensor
|
||||
:rtype: torch.Tensor
|
||||
"""
|
||||
if d is None:
|
||||
d = input_.labels
|
||||
|
||||
if components is None:
|
||||
components = output_.labels
|
||||
|
||||
if not isinstance(components, list):
|
||||
components = [components]
|
||||
|
||||
if not isinstance(d, list):
|
||||
d = [d]
|
||||
|
||||
tmp = (
|
||||
grad(output_, input_, components, d)
|
||||
.reshape(-1, len(components), len(d))
|
||||
.transpose(0, 1)
|
||||
components, d = _check_values(
|
||||
output_=output_, input_=input_, components=components, d=d
|
||||
)
|
||||
|
||||
tmp *= output_.extract(velocity_field)
|
||||
return tmp.sum(dim=2).T
|
||||
# Check if velocity field is present in the output labels
|
||||
if velocity_field not in output_.labels:
|
||||
raise RuntimeError(
|
||||
f"Velocity {velocity_field} is not present in the output labels."
|
||||
)
|
||||
|
||||
return fast_advection(
|
||||
output_=output_,
|
||||
input_=input_,
|
||||
velocity_field=velocity_field,
|
||||
components=components,
|
||||
d=d,
|
||||
)
|
||||
|
||||
Reference in New Issue
Block a user