mirror of
https://github.com/finegrain-ai/refiners.git
synced 2024-11-09 23:12:02 +00:00
162 lines
5.8 KiB
Python
162 lines
5.8 KiB
Python
import argparse
|
|
from pathlib import Path
|
|
|
|
import torch
|
|
|
|
from refiners.fluxion.utils import load_tensors, save_to_safetensors
|
|
|
|
|
|
def convert_dinov2_facebook(weights: dict[str, torch.Tensor]) -> None:
|
|
"""Convert a DINOv2 weights from facebook to refiners."""
|
|
# get depth from "blocks" keys
|
|
depth = max([int(k.split(".")[1]) for k in weights.keys() if k.startswith("blocks.")]) + 1
|
|
|
|
# only needed when pre-training
|
|
del weights["mask_token"]
|
|
|
|
# squeeze cls_token and position_embeddings
|
|
weights["cls_token"] = weights["cls_token"].squeeze(0)
|
|
weights["pos_embed"] = weights["pos_embed"].squeeze(0)
|
|
|
|
rename_keys: list[tuple[str, str]] = [
|
|
("cls_token", "Concatenate.ClassToken.Parameter.weight"),
|
|
("pos_embed", "PositionalEncoder.Parameter.weight"),
|
|
("patch_embed.proj.weight", "Concatenate.PatchEncoder.Conv2d.weight"),
|
|
("patch_embed.proj.bias", "Concatenate.PatchEncoder.Conv2d.bias"),
|
|
("norm.weight", "LayerNorm.weight"),
|
|
("norm.bias", "LayerNorm.bias"),
|
|
]
|
|
for i in range(depth):
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.norm1.weight",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_1.LayerNorm.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.norm1.bias",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_1.LayerNorm.bias",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.attn.proj.weight",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Linear.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.attn.proj.bias",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Linear.bias",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.ls1.gamma",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_1.LayerScale.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.norm2.weight",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.LayerNorm.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.norm2.bias",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.LayerNorm.bias",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.mlp.fc1.weight",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.FeedForward.Linear_1.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.mlp.fc1.bias",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.FeedForward.Linear_1.bias",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.mlp.fc2.weight",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.FeedForward.Linear_2.weight",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.mlp.fc2.bias",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.FeedForward.Linear_2.bias",
|
|
),
|
|
)
|
|
rename_keys.append(
|
|
(
|
|
f"blocks.{i}.ls2.gamma",
|
|
f"Transformer.TransformerLayer_{i+1}.Residual_2.LayerScale.weight",
|
|
),
|
|
)
|
|
|
|
if "register_tokens" in weights:
|
|
weights["register_tokens"] = weights["register_tokens"].squeeze(0)
|
|
rename_keys.append(("register_tokens", "Registers.Parameter.weight"))
|
|
|
|
# rename keys
|
|
for old_key, new_key in rename_keys:
|
|
weights[new_key] = weights.pop(old_key)
|
|
|
|
# split the qkv weights and biases
|
|
for i in range(depth):
|
|
qkv_weight = weights.pop(f"blocks.{i}.attn.qkv.weight")
|
|
q_weight, k_weight, v_weight = qkv_weight.chunk(3, dim=0)
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_1.weight"] = q_weight
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_2.weight"] = k_weight
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_3.weight"] = v_weight
|
|
|
|
qkv_bias = weights.pop(f"blocks.{i}.attn.qkv.bias")
|
|
q_bias, k_bias, v_bias = qkv_bias.chunk(3, dim=0)
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_1.bias"] = q_bias
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_2.bias"] = k_bias
|
|
weights[f"Transformer.TransformerLayer_{i+1}.Residual_1.SelfAttention.Distribute.Linear_3.bias"] = v_bias
|
|
|
|
|
|
def main() -> None:
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument(
|
|
"--from",
|
|
type=str,
|
|
required=True,
|
|
dest="source_path",
|
|
help=(
|
|
"Official checkpoint from https://github.com/facebookresearch/dinov2"
|
|
" e.g. /path/to/dinov2_vits14_pretrain.pth"
|
|
),
|
|
)
|
|
parser.add_argument(
|
|
"--to",
|
|
type=str,
|
|
dest="output_path",
|
|
default=None,
|
|
help=(
|
|
"Path to save the converted model. If not specified, the output path will be the source path with the"
|
|
" extension changed to .safetensors."
|
|
),
|
|
)
|
|
parser.add_argument("--half", action="store_true", dest="half")
|
|
args = parser.parse_args()
|
|
|
|
weights = load_tensors(args.source_path)
|
|
convert_dinov2_facebook(weights)
|
|
if args.half:
|
|
weights = {key: value.half() for key, value in weights.items()}
|
|
if args.output_path is None:
|
|
args.output_path = f"{Path(args.source_path).stem}.safetensors"
|
|
save_to_safetensors(path=args.output_path, tensors=weights)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|