|
|
|
--- |
|
|
|
license: apache-2.0 |
|
tags: |
|
- generated-by-script |
|
- peft |
|
- image-captioning |
|
base_model: [] |
|
|
|
|
|
--- |
|
|
|
# Model: ashimdahal/microsoft-git-base_microsoft-git-base |
|
|
|
This repository contains model artifacts for a run named `microsoft-git-base_microsoft-git-base`, likely a PEFT adapter. |
|
|
|
## Training Source |
|
This model was trained as part of the project/codebase available at: |
|
https://github.com/ashimdahal/captioning_image/blob/main |
|
|
|
## Base Model Information (Heuristic) |
|
* **Processor/Vision Encoder (Guessed):** `microsoft/git-base` |
|
* **Decoder/Language Model (Guessed):** `microsoft/git-base` |
|
|
|
**⚠️ Important:** The `base_model` tag in the metadata above is initially empty. The models listed here are *heuristic guesses* based on the training directory name (`microsoft-git-base_microsoft-git-base`). Please verify these against your training configuration and update the `base_model:` list in the YAML metadata block at the top of this README with the correct Hugging Face model identifiers. |
|
|
|
## How to Use (Example with PEFT) |
|
|
|
```python |
|
from transformers import AutoProcessor, AutoModelForVision2Seq, Blip2ForConditionalGeneration # Or other relevant classes |
|
from peft import PeftModel, PeftConfig |
|
import torch |
|
|
|
# --- Configuration --- |
|
# 1. Specify the EXACT base model identifiers used during training |
|
base_processor_id = "microsoft/git-base" # <-- Replace with correct HF ID |
|
base_model_id = "microsoft/git-base" # <-- Replace with correct HF ID (e.g., Salesforce/blip2-opt-2.7b) |
|
|
|
# 2. Specify the PEFT adapter repository ID (this repo) |
|
adapter_repo_id = "ashimdahal/microsoft-git-base_microsoft-git-base" |
|
|
|
# --- Load Base Model and Processor --- |
|
processor = AutoProcessor.from_pretrained(base_processor_id) |
|
|
|
# Load the base model (ensure it matches the type used for training) |
|
# Example for BLIP-2 OPT: |
|
base_model = Blip2ForConditionalGeneration.from_pretrained( |
|
base_model_id, |
|
torch_dtype=torch.float16 # Or torch.bfloat16 or float32, match training/inference needs |
|
) |
|
# Or for other model types: |
|
base_model = AutoModelForVision2Seq.from_pretrained(base_model_id, torch_dtype=torch.float16) |
|
base_model = AutoModelForCausalLM |
|
...... |
|
|
|
# --- Load PEFT Adapter --- |
|
# Load the adapter config and merge the adapter weights into the base model |
|
model = PeftModel.from_pretrained(base_model, adapter_repo_id) |
|
model = model.merge_and_unload() # Merge weights for inference (optional but often recommended) |
|
model.eval() # Set model to evaluation mode |
|
|
|
# --- Inference Example --- |
|
device = "cuda" if torch.cuda.is_available() else "cpu" |
|
model.to(device) |
|
|
|
image = ... # Load your image (e.g., using PIL) |
|
text = "a photo of" # Optional prompt start |
|
|
|
inputs = processor(images=image, text=text, return_tensors="pt").to(device, torch.float16) # Match model dtype |
|
|
|
generated_ids = model.generate(**inputs, max_new_tokens=50) |
|
generated_text = processor.batch_decode(generated_ids, skip_special_tokens=True)[0].strip() |
|
print(f"Generated Caption: {{generated_text}}") |
|
``` |
|
|
|
*More model-specific documentation, evaluation results, and usage examples should be added here.* |
|
|