Skip to content

Commit

Permalink
v0.0.13
Browse files Browse the repository at this point in the history
  • Loading branch information
Josh-XT committed Oct 4, 2023
1 parent 3a992ee commit af6c744
Show file tree
Hide file tree
Showing 2 changed files with 9 additions and 7 deletions.
14 changes: 8 additions & 6 deletions local_llm/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -230,14 +230,16 @@ def __init__(
self.params["model_path"] = get_model(
model_name=self.model_name, models_dir=models_dir
)
model_max_tokens = get_max_tokens(
model_name=self.model_name, models_dir=models_dir
)
self.prompt_template = get_prompt(
model_name=self.model_name, models_dir=models_dir
)
else:
self.params["model_path"] = ""
model_max_tokens = get_max_tokens(
model_name=self.model_name, models_dir=models_dir
)
self.prompt_template = get_prompt(
model_name=self.model_name, models_dir=models_dir
)
model_max_tokens = 8192
self.prompt_template = "{system_message}\n\n{prompt}"
try:
self.max_tokens = (
int(max_tokens)
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@

setup(
name="local-llm",
version="0.0.12",
version="0.0.13",
description="Local-LLM is a llama.cpp server in Docker with OpenAI Style Endpoints.",
long_description=long_description,
long_description_content_type="text/markdown",
Expand Down

0 comments on commit af6c744

Please sign in to comment.