make clip g use quick gelu and pad_token_id 0

This commit is contained in:
limiteinductive 2023-08-15 16:44:45 +02:00 committed by Benjamin Trom
parent 6594502c11
commit b8e7179447

View file

@ -247,11 +247,14 @@ class CLIPTextEncoderG(CLIPTextEncoder):
"""
def __init__(self, device: Device | str | None = None, dtype: DType | None = None) -> None:
tokenizer = CLIPTokenizer(pad_token_id=0)
super().__init__(
embedding_dim=1280,
num_layers=32,
num_attention_heads=20,
feedforward_dim=5120,
use_quick_gelu=True,
tokenizer=tokenizer,
device=device,
dtype=dtype,
)