mirror of
https://github.com/finegrain-ai/refiners.git
synced 2024-11-14 00:58:13 +00:00
add tests based on repr for inject / eject
This commit is contained in:
parent
0e77ef1720
commit
d185711bc5
|
@ -11,6 +11,8 @@ from refiners.foundationals.clip.tokenizer import CLIPTokenizer
|
||||||
@pytest.mark.parametrize("k_encoder", [CLIPTextEncoderL])
|
@pytest.mark.parametrize("k_encoder", [CLIPTextEncoderL])
|
||||||
def test_inject_eject(k_encoder: type[CLIPTextEncoder], test_device: torch.device):
|
def test_inject_eject(k_encoder: type[CLIPTextEncoder], test_device: torch.device):
|
||||||
encoder = k_encoder(device=test_device)
|
encoder = k_encoder(device=test_device)
|
||||||
|
initial_repr = repr(encoder)
|
||||||
|
|
||||||
extender = ConceptExtender(encoder)
|
extender = ConceptExtender(encoder)
|
||||||
|
|
||||||
cat_embedding = torch.randn((encoder.embedding_dim,), device=test_device)
|
cat_embedding = torch.randn((encoder.embedding_dim,), device=test_device)
|
||||||
|
@ -18,7 +20,9 @@ def test_inject_eject(k_encoder: type[CLIPTextEncoder], test_device: torch.devic
|
||||||
|
|
||||||
extender_2 = ConceptExtender(encoder)
|
extender_2 = ConceptExtender(encoder)
|
||||||
|
|
||||||
|
assert repr(encoder) == initial_repr
|
||||||
extender.inject()
|
extender.inject()
|
||||||
|
assert repr(encoder) != initial_repr
|
||||||
|
|
||||||
with pytest.raises(AssertionError) as no_nesting:
|
with pytest.raises(AssertionError) as no_nesting:
|
||||||
extender_2.inject()
|
extender_2.inject()
|
||||||
|
@ -34,6 +38,7 @@ def test_inject_eject(k_encoder: type[CLIPTextEncoder], test_device: torch.devic
|
||||||
|
|
||||||
extender_2.inject().eject()
|
extender_2.inject().eject()
|
||||||
ConceptExtender(encoder) # no exception
|
ConceptExtender(encoder) # no exception
|
||||||
|
assert repr(encoder) == initial_repr
|
||||||
|
|
||||||
tokenizer = encoder.ensure_find(CLIPTokenizer)
|
tokenizer = encoder.ensure_find(CLIPTokenizer)
|
||||||
assert len(tokenizer.encode("<token1>")) > 3
|
assert len(tokenizer.encode("<token1>")) > 3
|
||||||
|
|
37
tests/adapters/test_ip_adapter.py
Normal file
37
tests/adapters/test_ip_adapter.py
Normal file
|
@ -0,0 +1,37 @@
|
||||||
|
from typing import overload
|
||||||
|
|
||||||
|
import pytest
|
||||||
|
import torch
|
||||||
|
|
||||||
|
from refiners.fluxion.utils import no_grad
|
||||||
|
from refiners.foundationals.latent_diffusion import SD1IPAdapter, SD1UNet, SDXLIPAdapter, SDXLUNet
|
||||||
|
|
||||||
|
|
||||||
|
@overload
|
||||||
|
def new_adapter(target: SD1UNet) -> SD1IPAdapter:
|
||||||
|
...
|
||||||
|
|
||||||
|
|
||||||
|
@overload
|
||||||
|
def new_adapter(target: SDXLUNet) -> SDXLIPAdapter:
|
||||||
|
...
|
||||||
|
|
||||||
|
|
||||||
|
def new_adapter(target: SD1UNet | SDXLUNet) -> SD1IPAdapter | SDXLIPAdapter:
|
||||||
|
if isinstance(target, SD1UNet):
|
||||||
|
return SD1IPAdapter(target=target)
|
||||||
|
else:
|
||||||
|
return SDXLIPAdapter(target=target)
|
||||||
|
|
||||||
|
|
||||||
|
@no_grad()
|
||||||
|
@pytest.mark.parametrize("k_unet", [SD1UNet, SDXLUNet])
|
||||||
|
def test_inject_eject(k_unet: type[SD1UNet] | type[SDXLUNet], test_device: torch.device):
|
||||||
|
unet = k_unet(in_channels=4, device=test_device)
|
||||||
|
initial_repr = repr(unet)
|
||||||
|
adapter = new_adapter(unet)
|
||||||
|
assert repr(unet) == initial_repr
|
||||||
|
adapter.inject()
|
||||||
|
assert repr(unet) != initial_repr
|
||||||
|
adapter.eject()
|
||||||
|
assert repr(unet) == initial_repr
|
|
@ -29,7 +29,11 @@ def new_adapter(target: SD1UNet | SDXLUNet, name: str) -> SD1T2IAdapter | SDXLT2
|
||||||
@pytest.mark.parametrize("k_unet", [SD1UNet, SDXLUNet])
|
@pytest.mark.parametrize("k_unet", [SD1UNet, SDXLUNet])
|
||||||
def test_inject_eject(k_unet: type[SD1UNet] | type[SDXLUNet], test_device: torch.device):
|
def test_inject_eject(k_unet: type[SD1UNet] | type[SDXLUNet], test_device: torch.device):
|
||||||
unet = k_unet(in_channels=4, device=test_device)
|
unet = k_unet(in_channels=4, device=test_device)
|
||||||
adapter_1 = new_adapter(unet, "adapter 1").inject()
|
initial_repr = repr(unet)
|
||||||
|
adapter_1 = new_adapter(unet, "adapter 1")
|
||||||
|
assert repr(unet) == initial_repr
|
||||||
|
adapter_1.inject()
|
||||||
|
assert repr(unet) != initial_repr
|
||||||
|
|
||||||
with pytest.raises(AssertionError) as already_injected_error:
|
with pytest.raises(AssertionError) as already_injected_error:
|
||||||
new_adapter(unet, "adapter 1").inject()
|
new_adapter(unet, "adapter 1").inject()
|
||||||
|
@ -50,3 +54,4 @@ def test_inject_eject(k_unet: type[SD1UNet] | type[SDXLUNet], test_device: torch
|
||||||
|
|
||||||
assert unet.parent is None
|
assert unet.parent is None
|
||||||
assert unet.find(T2IFeatures) is None
|
assert unet.find(T2IFeatures) is None
|
||||||
|
assert repr(unet) == initial_repr
|
||||||
|
|
Loading…
Reference in a new issue