File size: 962 Bytes
493728d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
"""The Long-Context Cerebras Planning and Optimization (LongCePO) Method
LongCePO is an inference-time computation method designed to provide LLMs with the capability to work with infinite context such as external knowledge bases that can run into millions of tokens. We achieve this goal through a combination of multiple strategies including planning (query decomposition) and divide-and-conquer long-context processing. This approach enables to use a limited context window (e.g. 8K) and outperform full-context processing with the same base model in many question-answering tasks.
If you have any questions or want to contribute, please reach out to us on [cerebras.ai/discord](https://cerebras.ai/discord).
"""
from typing import Tuple
from .longcepo.main import run_longcepo
SLUG = "longcepo"
def run(system_prompt: str, initial_query: str, client, model: str) -> Tuple[str, int]:
return run_longcepo(system_prompt, initial_query, client, model)
|