Spaces:
Paused
Paused
File size: 796 Bytes
7396aab f414499 7396aab f414499 7396aab |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
import torch
from transformers import PretrainedConfig, BitsAndBytesConfig
import math
from typing import Optional
class VisionProjectorConfig(PretrainedConfig):
def __init__(
self,
input_dim=768,
hidden_dim=256,
num_tokens=1,
output_dim=2560,
**kwargs
):
#super.__init__(**kwargs)
self.input_dim = input_dim
self.hidden_dim = hidden_dim
self.output_dim = output_dim
self.num_tokens = num_tokens
self.kwargs = kwargs
app_config = dict(
max_seqlen=512,
max_caption_len=100,
data_dir='../data',
output_dir="./results",
vision_model=True,
vision_projector_file='models/vision_projector/vp_ckpt_0.pth',
phi_adapter_dir='models/phi_adapter'
) |