Skip to content
Snippets Groups Projects
Commit b4f9fd21 authored by Max Kimmich's avatar Max Kimmich Committed by Grießhaber Daniel
Browse files

Add CLI option for maximum number of tokens being generated

parent 3ae988c3
No related branches found
No related tags found
2 merge requests!2remove is_chat argument,!1Refactor models
......@@ -72,7 +72,11 @@ class LLMModel(ABC):
stop: str = None,
history: list[dict[str, str]] | None = None,
**kwargs: Any,
) -> tuple[str, list[dict[str, str]], ModelUsage]:
) -> tuple[str, ModelUsage]:
if chat is None:
chat = self.chat
max_tokens = kwargs.pop("max_tokens", self.options.max_tokens)
# create prompt
prompt = prompt_prefix + prompt + prompt_suffix + prompt_appendix
messages = [self._get_user_message(prompt)]
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment