vision_transformer_cloak
Module containing the VisionTransformerCloak class, which implements an
input dependent Stained Glass Transform for computer vision problems using
DinoV2 as the feature extractor.
Classes:
| Name | Description |
|---|---|
VisionTransformerCloak |
Facade for the input dependent Stained Glass Transform for computer vision problems using DinoV2 as the feature extractor. |
VisionTransformerEstimator |
Initialize an input dependent noise estimator for computer vision problems using DinoV2 as the feature extractor. |
VisionTransformerCloak
¶
Bases: BaseNoiseLayer[DinoV2, CloakStandardDeviationParameterization, BatchwisePercentMasker | None]
Facade for the input dependent Stained Glass Transform for computer vision problems using DinoV2 as the feature extractor.
Added in version v3.8.0.
Methods:
| Name | Description |
|---|---|
__call__ |
Transform the input data. |
__getstate__ |
Prepare a JSON-serializable copy of the noise layer's state. |
__init__ |
Initialize a |
__setstate__ |
Set the state of the object. |
forward |
Compute the transformed output from the input and noise mask. |
get_applied_transform_components_factory |
Create a function that returns the elements of the transform components ( |
get_transformed_output_factory |
Create a function that returns the transformed output from the most recent forward pass. |
initial_seed |
Return the initial seed of the CPU device's random number generator. |
manual_seed |
Seed each of the random number generators. |
reset_parameters |
Reinitialize parameters and buffers. |
seed |
Seed each of the random number generators using a non-deterministic random number. |
__call__
¶
Transform the input data.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
Tensor
|
The input to transform. |
required |
|
Tensor | None
|
An optional mask that selects the elements of |
None
|
|
Any
|
Additional keyword arguments to the estimator modules. |
required |
__getstate__
¶
Prepare a JSON-serializable copy of the noise layer's state.
Returns:
| Type | Description |
|---|---|
dict[str, Any]
|
A dictionary containing the configuration of the noise layer, including its type string, the state dict, and the generator |
dict[str, Any]
|
states if they exist. |
__init__
¶
__init__(
scale: tuple[float, float] = (0.0001, 2.0),
shallow: float = 1.0,
value_range: (
tuple[float | None, float | None] | None
) = None,
seed: int | None = None,
) -> None
Initialize a VisionTransformerCloak class.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
tuple[float, float]
|
The min and max scale for the standard deviation parameterization. |
(0.0001, 2.0)
|
|
float
|
A hyperparameter smoothing the tanh parameterizing the std estimator. |
1.0
|
|
tuple[float | None, float | None] | None
|
The range of features to limit the output of the SGT to. |
None
|
|
int | None
|
A seed for the RNG. |
None
|
__setstate__
¶
__setstate__(
state: dict[str, Any],
trust_remote_code: bool = False,
third_party_model_path: (
str | PathLike[str] | None
) = None,
) -> None
Set the state of the object.
state_dict and _generators are both optional keys, and will be restored if they exist in the state.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
dict[str, Any]
|
The state to set. |
required |
|
bool
|
Whether to trust remote code when loading from HuggingFace Hub. |
False
|
|
str | PathLike[str] | None
|
The path or huggingface reference to a third-party model to load. This is useful when loading SGTs whose internal structure depends on transformers which are not importable directly through transformers, but are present on the Hugging Face Hub. |
None
|
forward
¶
Compute the transformed output from the input and noise mask.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
Tensor
|
The input to transform. |
required |
|
Tensor | None
|
An optional noise mask to apply to the output. |
None
|
|
Any
|
Additional keyword arguments to pass to the estimators. |
required |
Returns:
| Type | Description |
|---|---|
torch.Tensor
|
The Stained Glass transformed output. |
get_applied_transform_components_factory
¶
Create a function that returns the elements of the transform components ('mean' and 'std') applied during the most recent
forward pass.
Specifically, the applied elements are those selected by the noise mask (if supplied) and standard deviation mask (if
std_estimator.masker is not None). If no masks are used, all elements are returned.
The applied transform components are returned flattened.
This function is intended to be used to log histograms of the transform components.
Returns:
| Type | Description |
|---|---|
Callable[[], dict[str, torch.Tensor]]
|
A function that returns the the elements of the transform components applied during the most recent forward pass. |
Examples:
>>> from torch import nn
>>> from stainedglass_core import model as sg_model, noise_layer as sg_noise_layer
>>> base_model = nn.Linear(20, 2)
>>> noisy_model = sg_model.NoisyModel(
... sg_noise_layer.CloakNoiseLayer1,
... base_model,
... target_parameter="input",
... )
>>> get_applied_transform_components = (
... noisy_model.noise_layer.get_applied_transform_components_factory()
... )
>>> input = torch.ones(1, 20)
>>> noise_mask = torch.tensor(5 * [False] + 15 * [True])
>>> output = noisy_model(input, noise_mask=noise_mask)
>>> applied_transform_components = get_applied_transform_components()
>>> applied_transform_components
{'mean': tensor(...), 'std': tensor(...)}
>>> {
... component_name: component.shape
... for component_name, component in applied_transform_components.items()
... }
{'mean': torch.Size([15]), 'std': torch.Size([15])}
get_transformed_output_factory
¶
Create a function that returns the transformed output from the most recent forward pass.
If super batching is active, only the transformed half of the super batch output is returned.
Returns:
| Type | Description |
|---|---|
Callable[[], torch.Tensor]
|
A function that returns the transformed output from the most recent forward pass. |
Examples:
>>> from stainedglass_core import noise_layer as sg_noise_layer
>>> noise_layer = sg_noise_layer.CloakNoiseLayer1()
>>> get_transformed_output = noise_layer.get_transformed_output_factory()
>>> input = torch.ones(2, 3, 32, 32)
>>> output = noise_layer(input)
>>> transformed_output = get_transformed_output()
>>> assert output.equal(transformed_output)
initial_seed
¶
Return the initial seed of the CPU device's random number generator.
manual_seed
¶
manual_seed(
seed: int | None, rank_dependent: bool = True
) -> None
Seed each of the random number generators.
Setting seed to None will destroy any existing generators.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
int | None
|
The seed to set. |
required |
|
bool
|
Whether to add the distributed rank to the seed to ensure that each process samples different noise. |
True
|
reset_parameters
¶
Reinitialize parameters and buffers.
This method is useful for initializing tensors created on the meta device.
seed
¶
Seed each of the random number generators using a non-deterministic random number.
VisionTransformerEstimator
¶
Bases: Estimator[DinoV2, OptionalParameterizationT, OptionalMaskerT_contra]
Initialize an input dependent noise estimator for computer vision problems using DinoV2 as the feature extractor.
Added in version v3.8.0.
Methods:
| Name | Description |
|---|---|
__call__ |
Estimate noise components from input values. |
__init__ |
Initialize an input dependent estimator which uses DinoV2 as the feature extractor. |
forward |
Compute the forward pass of the estimator block. |
reset_parameters |
Reinitialize parameters and buffers. |
__call__
¶
__call__(
input: Tensor,
noise_mask: Tensor | None = None,
**kwargs: Any
) -> tuple[torch.Tensor, torch.Tensor | None]
__call__(
input: Tensor,
noise_mask: Tensor | None = None,
**kwargs: Any
) -> tuple[torch.Tensor, torch.Tensor | None]
__call__(
input: Tensor,
noise_mask: Tensor | None = None,
**kwargs: Any
) -> tuple[torch.Tensor, torch.Tensor | None]
__call__(
input: Tensor,
noise_mask: Tensor | None = None,
**kwargs: Any
) -> tuple[torch.Tensor, torch.Tensor | None]
Estimate noise components from input values.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
Tensor
|
The tensor to estimate noise components from. |
required |
|
Tensor | None
|
An optional mask tensor to use to select a subset of the elements of the estimated standard deviations for computing
a mask to apply to the input. If |
None
|
|
Any
|
Additional keyword arguments to |
required |
Returns:
| Type | Description |
|---|---|
tuple[torch.Tensor, torch.Tensor | None]
|
A tuple containing the estimated noise components and an input mask if there is a |
__init__
¶
__init__(
module: DinoV2,
target_parameter: str = "tensor_list",
parameterization: OptionalParameterizationT = None,
masker: OptionalMaskerT_contra = None,
) -> None
Initialize an input dependent estimator which uses DinoV2 as the feature extractor.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
DinoV2
|
The DinoV2 module to use as the feature extractor. |
required |
|
str
|
The target parameter to apply the stained glass transform to. |
'tensor_list'
|
|
OptionalParameterizationT
|
The optional parameterization to apply the output to. |
None
|
|
OptionalMaskerT_contra
|
The optional masker to apply to the output. |
None
|
Raises:
| Type | Description |
|---|---|
TypeError
|
If the encoder module of the given DinoV2 module's backbone's encoder is not a DinoV2. |
forward
¶
forward(
input: Tensor,
noise_mask: Tensor | None = None,
**kwargs: Any
) -> tuple[torch.Tensor, torch.Tensor | None]
Compute the forward pass of the estimator block.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
|
Tensor
|
The input to the estimator block. |
required |
|
Tensor | None
|
The optional noise mask. This is kept around for legacy support and is not meaningfully used in computer vision problems. |
None
|
|
Any
|
Additional keyword arguments to pass to the underlying estimator module. |
required |
Returns:
| Type | Description |
|---|---|
tuple[torch.Tensor, torch.Tensor | None]
|
The estimated output depending on the input and an optional mask. |
reset_parameters
¶
Reinitialize parameters and buffers.
This method is useful for initializing tensors created on the meta device.