Spaces:
Running
Running
Update models.py
Browse files
models.py
CHANGED
@@ -11,38 +11,34 @@ import gradio_helpers
|
|
11 |
import paligemma_bv
|
12 |
|
13 |
|
14 |
-
ORGANIZATION = '
|
15 |
BASE_MODELS = [
|
16 |
-
('paligemma-3b-mix-224-jax', 'paligemma-3b-mix-224'),
|
17 |
-
('paligemma-3b-mix-448-jax', 'paligemma-3b-mix-448'),
|
|
|
18 |
]
|
19 |
MODELS = {
|
20 |
-
**{
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
},
|
|
|
28 |
}
|
29 |
|
30 |
MODELS_INFO = {
|
31 |
-
'paligemma-3b-
|
32 |
'JAX/FLAX PaliGemma 3B weights, finetuned with 224x224 input images and 256 token input/output '
|
33 |
'text sequences on a mixture of downstream academic datasets. The models are available in float32, '
|
34 |
'bfloat16 and float16 format for research purposes only.'
|
35 |
),
|
36 |
-
'paligemma-3b-mix-448': (
|
37 |
-
'JAX/FLAX PaliGemma 3B weights, finetuned with 448x448 input images and 512 token input/output '
|
38 |
-
'text sequences on a mixture of downstream academic datasets. The models are available in float32, '
|
39 |
-
'bfloat16 and float16 format for research purposes only.'
|
40 |
-
),
|
41 |
}
|
42 |
|
43 |
MODELS_RES_SEQ = {
|
44 |
-
'paligemma-3b-
|
45 |
-
'paligemma-3b-mix-448': (448, 512),
|
46 |
}
|
47 |
|
48 |
# "CPU basic" has 16G RAM, "T4 small" has 15 GB RAM.
|
@@ -53,7 +49,7 @@ MAX_RAM_CACHE = int(float(os.environ.get('RAM_CACHE_GB', '0')) * 1e9)
|
|
53 |
config = paligemma_bv.PaligemmaConfig(
|
54 |
ckpt='', # will be set below
|
55 |
res=224,
|
56 |
-
text_len=
|
57 |
tokenizer='gemma(tokensets=("loc", "seg"))',
|
58 |
vocab_size=256_000 + 1024 + 128,
|
59 |
)
|
|
|
11 |
import paligemma_bv
|
12 |
|
13 |
|
14 |
+
ORGANIZATION = 'Jegree'
|
15 |
BASE_MODELS = [
|
16 |
+
# ('paligemma-3b-mix-224-jax', 'paligemma-3b-mix-224'),
|
17 |
+
# ('paligemma-3b-mix-448-jax', 'paligemma-3b-mix-448'),
|
18 |
+
('myPaligem', 'fine-tuned-paligemma-3b-pt-224')
|
19 |
]
|
20 |
MODELS = {
|
21 |
+
# **{
|
22 |
+
# model_name: (
|
23 |
+
# f'{ORGANIZATION}/{repo}',
|
24 |
+
# f'{model_name}.bf16.npz',
|
25 |
+
# 'bfloat16', # Model repo revision.
|
26 |
+
# )
|
27 |
+
# for repo, model_name in BASE_MODELS
|
28 |
+
# },
|
29 |
+
'fine-tuned-paligemma-3b-pt-224':('Jegree/myPaligem', 'fine-tuned-paligemma-3b-pt-224.f16.npz', 'main'),
|
30 |
}
|
31 |
|
32 |
MODELS_INFO = {
|
33 |
+
'fine-tuned-paligemma-3b-pt-224': (
|
34 |
'JAX/FLAX PaliGemma 3B weights, finetuned with 224x224 input images and 256 token input/output '
|
35 |
'text sequences on a mixture of downstream academic datasets. The models are available in float32, '
|
36 |
'bfloat16 and float16 format for research purposes only.'
|
37 |
),
|
|
|
|
|
|
|
|
|
|
|
38 |
}
|
39 |
|
40 |
MODELS_RES_SEQ = {
|
41 |
+
'fine-tuned-paligemma-3b-pt-224': (224, 128),
|
|
|
42 |
}
|
43 |
|
44 |
# "CPU basic" has 16G RAM, "T4 small" has 15 GB RAM.
|
|
|
49 |
config = paligemma_bv.PaligemmaConfig(
|
50 |
ckpt='', # will be set below
|
51 |
res=224,
|
52 |
+
text_len=128,
|
53 |
tokenizer='gemma(tokensets=("loc", "seg"))',
|
54 |
vocab_size=256_000 + 1024 + 128,
|
55 |
)
|