Skip to content

Commit

Permalink
add llama 405b
Browse files Browse the repository at this point in the history
  • Loading branch information
shivanker committed Jul 24, 2024
1 parent c02f869 commit 8f82358
Show file tree
Hide file tree
Showing 2 changed files with 15 additions and 11 deletions.
7 changes: 4 additions & 3 deletions lite_llms.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,8 @@ class TextModel(Enum):
CLAUDE_3_HAIKU = "claude-3-haiku-20240307"
GEMINI_15_PRO = "gemini/gemini-1.5-pro-latest"
GEMINI_15_FLASH = "gemini/gemini-1.5-flash-latest"
GROQ_LLAMA3_70B = "groq/llama3-70b-8192"
GROQ_LLAMA3_8B = "groq/llama3-8b-8192"
# GROQ_LLAMA3_70B = "groq/llama3-70b-8192"
# GROQ_LLAMA3_8B = "groq/llama3-8b-8192"
LLAMA3_70B = "fireworks_ai/llama-v3-70b-instruct"
LLAMA3_8B = "fireworks_ai/llama-v3-8b-instruct"
LLAMA31_405B = "fireworks_ai/llama-v3p1-405b-instruct"
LLAMA31_8B = "fireworks_ai/llama-v3p1-8b-instruct"
19 changes: 11 additions & 8 deletions session.py
Original file line number Diff line number Diff line change
Expand Up @@ -229,15 +229,18 @@ def process_command(self, text, say=lambda text: None):
elif cmd == "\\gpt4":
self.model = TextModel.GPT_4_TURBO
say(text="Model set to GPT-4.")
elif cmd in ["\\llama70b", "\\llama70", "\\llama"]:
elif cmd in ["\\llama", "\\llama31", "\\llama405", "\\llama405b"]:
self.model = TextModel.LLAMA31_405B
say(text="Model set to LLaMA-3.1 405B.")
elif cmd in ["\\llama70b", "\\llama70"]:
self.model = TextModel.LLAMA3_70B
say(text="Model set to LLaMA 3 70B.")
say(text="Model set to LLaMA-3 70B.")
elif cmd in ["\\llama8b", "\\llama8"]:
self.model = TextModel.LLAMA3_8B
say(text="Model set to LLaMA 3 8B.")
elif cmd in ["\\groq", "\\groq70", "\\groq70b"]:
self.model = TextModel.GROQ_LLAMA3_70B
say(text="Model set to LLaMA 3 70B (Groq).")
self.model = TextModel.LLAMA31_8B
say(text="Model set to LLaMA-3.1 8B.")
# elif cmd in ["\\groq", "\\groq70", "\\groq70b"]:
# self.model = TextModel.GROQ_LLAMA3_70B
# say(text="Model set to LLaMA 3 70B (Groq).")
elif cmd == "\\opus":
self.model = TextModel.CLAUDE_3_OPUS
say(text="Model set to Claude 3 Opus.")
Expand Down Expand Up @@ -270,7 +273,7 @@ def process_command(self, text, say=lambda text: None):
- \\gpt4o: Use GPT-4o (Omni) for future messages. Preserves the session so far.\n
- \\mini: Use GPT-4o Mini for future messages. Preserves the session so far.\n
- \\sonnet: Use Claude 3.5 Sonnet for future messages. Preserves the session so far.\n
- \\llama70: Use LLaMA-3-70B for future messages. Preserves the session so far.\n
- \\llama: Use LLaMA-3.1 405B for future messages. Preserves the session so far.\n
- \\gemini: Use Gemini 1.5 Pro for future messages. Preserves the session so far.\n
- \\flash: Use Gemini 1.5 Flash for future messages. Preserves the session so far.\n
- \\stream: Toggle streaming mode. In streaming mode, the bot will send you a message every time it generates a new token.\n
Expand Down

0 comments on commit 8f82358

Please sign in to comment.