Commit
·
eabfc8b
1
Parent(s):
b77b636
Update app.py
Browse files
app.py
CHANGED
@@ -324,7 +324,7 @@ title_markdown = """
|
|
324 |
# 🌋 LLaVA: Large Language and Vision Assistant
|
325 |
[[Project Page]](https://llava-vl.github.io) [[Paper]](https://arxiv.org/abs/2304.08485) [[Code]](https://github.com/haotian-liu/LLaVA) [[Model]](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md)
|
326 |
|
327 |
-
ONLY WORKS WITH GPU!
|
328 |
"""
|
329 |
|
330 |
tos_markdown = """
|
@@ -591,8 +591,7 @@ if __name__ == "__main__":
|
|
591 |
logger.info(f"args: {args}")
|
592 |
|
593 |
model_path = "liuhaotian/llava-v1.5-13b"
|
594 |
-
bits =
|
595 |
-
# set bits=4 for T4, bits=8 for A10G (24G), and bits=16 for A100 (40G)
|
596 |
|
597 |
preload_models(model_path)
|
598 |
|
|
|
324 |
# 🌋 LLaVA: Large Language and Vision Assistant
|
325 |
[[Project Page]](https://llava-vl.github.io) [[Paper]](https://arxiv.org/abs/2304.08485) [[Code]](https://github.com/haotian-liu/LLaVA) [[Model]](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md)
|
326 |
|
327 |
+
ONLY WORKS WITH GPU! 8-bit is used by default to load 13B model in A10G-Large (24G). Set environment `bits` to 4 to load it in T4-medium (15G).
|
328 |
"""
|
329 |
|
330 |
tos_markdown = """
|
|
|
591 |
logger.info(f"args: {args}")
|
592 |
|
593 |
model_path = "liuhaotian/llava-v1.5-13b"
|
594 |
+
bits = os.getenv("bits", 8)
|
|
|
595 |
|
596 |
preload_models(model_path)
|
597 |
|