Spaces:
Running
Running
# Copyright 2025 the LlamaFactory team. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
from abc import ABC, abstractmethod | |
from collections.abc import AsyncGenerator | |
from dataclasses import dataclass | |
from typing import TYPE_CHECKING, Any, Literal, Optional, Union | |
if TYPE_CHECKING: | |
from transformers import PreTrainedModel, PreTrainedTokenizer | |
from vllm import AsyncLLMEngine | |
from ..data import Template | |
from ..data.mm_plugin import AudioInput, ImageInput, VideoInput | |
from ..extras.constants import EngineName | |
from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments | |
class Response: | |
response_text: str | |
response_length: int | |
prompt_length: int | |
finish_reason: Literal["stop", "length"] | |
class BaseEngine(ABC): | |
r"""Base class for inference engine of chat models. | |
Must implements async methods: chat(), stream_chat() and get_scores(). | |
""" | |
name: "EngineName" | |
model: Union["PreTrainedModel", "AsyncLLMEngine"] | |
tokenizer: "PreTrainedTokenizer" | |
can_generate: bool | |
template: "Template" | |
generating_args: dict[str, Any] | |
def __init__( | |
self, | |
model_args: "ModelArguments", | |
data_args: "DataArguments", | |
finetuning_args: "FinetuningArguments", | |
generating_args: "GeneratingArguments", | |
) -> None: | |
r"""Initialize an inference engine.""" | |
... | |
async def chat( | |
self, | |
messages: list[dict[str, str]], | |
system: Optional[str] = None, | |
tools: Optional[str] = None, | |
images: Optional[list["ImageInput"]] = None, | |
videos: Optional[list["VideoInput"]] = None, | |
audios: Optional[list["AudioInput"]] = None, | |
**input_kwargs, | |
) -> list["Response"]: | |
r"""Get a list of responses of the chat model.""" | |
... | |
async def stream_chat( | |
self, | |
messages: list[dict[str, str]], | |
system: Optional[str] = None, | |
tools: Optional[str] = None, | |
images: Optional[list["ImageInput"]] = None, | |
videos: Optional[list["VideoInput"]] = None, | |
audios: Optional[list["AudioInput"]] = None, | |
**input_kwargs, | |
) -> AsyncGenerator[str, None]: | |
r"""Get the response token-by-token of the chat model.""" | |
... | |
async def get_scores( | |
self, | |
batch_input: list[str], | |
**input_kwargs, | |
) -> list[float]: | |
r"""Get a list of scores of the reward model.""" | |
... | |