# Tensor Utils¶

Utils for generating stats from torch tensors.

opacus.utils.tensor_utils.calc_sample_norms(named_params, *, flat=True)[source]

Calculates the norm of the given tensors for each sample.

This function calculates the overall norm of the given tensors for each sample, assuming each batch’s dim is zero.

Parameters:
Return type:
Returns:

A list of tensor norms where length of the list is the number of layers

Example

>>> t1 = torch.rand((2, 5))
>>> t2 = torch.rand((2, 5))
>>> norms = calc_sample_norms([("1", t1), ("2", t2)])
>>> norms, norms.shape
([tensor([...])], torch.Size())

opacus.utils.tensor_utils.calc_sample_norms_one_layer(param)[source]

Calculates the norm of the given tensor (a single parameter) for each sample.

This function calculates the overall norm of the given tensor for each sample, assuming each batch’s dim is zero.

It is equivalent to: calc_sample_norms(named_params=((None, param),))

Parameters:

param (Tensor) – A tensor of shape [B, ...] where B is the size of the batch and is the 0th dimension.

Return type:

Tensor

Returns:

A tensor of norms

Example

>>> t1 = torch.rand((2, 5))
>>> norms = calc_sample_norms_one_layer(t1)
>>> norms, norms.shape
(tensor([...]), torch.Size())

opacus.utils.tensor_utils.filter_dilated_rows(tensor, dilation, dilated_kernel_size, kernel_size)[source]

A helper function that removes extra rows created during the process of implementing dilation.

Parameters:
Returns:

A tensor of shape (B, C, D_out, H_out, W_out, kernel_size, kernel_size, kernel_size) For D_out, H_out, W_out definitions see torch.nn.Unfold.

Example

>>> tensor = torch.zeros([1, 1, 3, 3, 3, 5, 5, 5])
>>> dilation = (2, 2, 2)
>>> dilated_kernel_size = (5, 5, 5)
>>> kernel_size = (3, 3, 3)
>>> filter_dilated_rows(tensor, dilation, dilated_kernel_size, kernel_size).shape
torch.Size([1, 1, 3, 3, 3, 3, 3, 3])

opacus.utils.tensor_utils.sum_over_all_but_batch_and_last_n(tensor, n_dims)[source]

Calculates the sum over all dimensions, except the first (batch dimension), and excluding the last n_dims.

This function will ignore the first dimension, and it will not aggregate over the last n_dims dimensions.

Parameters:
Return type:

Tensor

Returns:

A tensor of shape (B, ..., X[n_dims-1])

Example

>>> tensor = torch.ones(1, 2, 3, 4, 5)
>>> sum_over_all_but_batch_and_last_n(tensor, n_dims=2).shape
torch.Size([1, 4, 5])

opacus.utils.tensor_utils.unfold2d(input, *, kernel_size, padding, stride, dilation)[source]

See unfold()

opacus.utils.tensor_utils.unfold3d(tensor, *, kernel_size, padding=0, stride=1, dilation=1)[source]

Extracts sliding local blocks from an batched input tensor.

torch.nn.Unfold only supports 4D inputs (batched image-like tensors). This method implements the same action for 5D inputs

Parameters:
Returns:

A tensor of shape (B, C * np.product(kernel_size), L), where L - output spatial dimensions. See torch.nn.Unfold for more details

Example

>>> B, C, D, H, W = 3, 4, 5, 6, 7
>>> tensor = torch.arange(1, B*C*D*H*W + 1.).view(B, C, D, H, W)