|
import streamlit as st
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer,pipeline
|
|
import torch
|
|
|
|
st.title("quantization_Generator Fine tunning model")
|
|
|
|
|
|
model_dir = "quantization_model"
|
|
tokenizer = AutoTokenizer.from_pretrained(model_dir)
|
|
model = AutoModelForCausalLM.from_pretrained(model_dir)
|
|
|
|
|
|
code_generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
|
|
|
|
|
|
|
|
|
inputs_text=st.text_input("Please enter the text",value="def quicksort(arr):")
|
|
|
|
if st.button("submit"):
|
|
generated_code = code_generator(inputs_text, max_length=200, num_return_sequences=1)
|
|
|
|
st.write(generated_code[0]["generated_text"])
|
|
|