Spaces:
Running
Running
import os | |
from transformers import AutoTokenizer, AutoModelForCausalLM | |
import torch | |
from torch.nn.functional import cosine_similarity | |
from collections import Counter | |
import numpy as np | |
from device_manager import DeviceManager | |
class Gemma2BDependencies: | |
_instance = None | |
def __new__(cls): | |
if cls._instance is None: | |
cls._instance = super(Gemma2BDependencies, cls).__new__(cls) | |
hf_token = os.environ.get('HUGGINGFACE_TOKEN', None) | |
cls._instance.tokenizer = AutoTokenizer.from_pretrained("google/gemma-2b", token=hf_token) | |
cls._instance.model = AutoModelForCausalLM.from_pretrained("google/gemma-2b", token=hf_token) | |
cls._instance.device = DeviceManager() | |
cls._instance.model.to(cls._instance.device) | |
return cls._instance | |
def calculate_perplexity(self, text: str): | |
inputs = self.tokenizer(text, return_tensors="pt", | |
truncation=True, max_length=1024) | |
inputs = {k: v.to(self.device) for k, v in inputs.items()} | |
# Calculate the model's output | |
with torch.no_grad(): | |
outputs = self.model(**inputs, labels=inputs["input_ids"]) | |
loss = outputs.loss | |
perplexity = torch.exp(loss) | |
return perplexity.item() | |
def calculate_burstiness(self, text: str): | |
tokens = self.tokenizer.encode(text, add_special_tokens=False) | |
# Count token frequencies | |
frequency_counts = list(Counter(tokens).values()) | |
# Calculate variance and mean of frequencies | |
variance = np.var(frequency_counts) | |
mean = np.mean(frequency_counts) | |
# Compute Variance-to-Mean Ratio (VMR) for burstiness | |
vmr = variance / mean if mean > 0 else 0 | |
return vmr | |