Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Model wishlist #217

Closed
5 tasks done
nsarrazin opened this issue Apr 24, 2023 · 17 comments
Closed
5 tasks done

Model wishlist #217

nsarrazin opened this issue Apr 24, 2023 · 17 comments

Comments

@nsarrazin
Copy link
Member

nsarrazin commented Apr 24, 2023

Hey everyone!

Just opening an issue to track which models people would like to see supported with Serge.

  • Alpaca 7B, 13B & 30B
  • Alpaca 7B-Native
  • gpt4all
  • Vicuna 7B & 13B
  • Open Assistant 13B & 30B

Are there any others you would like to see ?

@morpheus2448

This comment was marked as abuse.

1 similar comment
@morpheus2448

This comment was marked as abuse.

@kolabearafk
Copy link

How about WizardLM (https://github.com/nlpxucan/WizardLM) ? Can this new model be supported in Serge please? I heard it's very fast and generates good results.

@CrazyBonze
Copy link

I was going to say WizardLM, its been getting some good reviews.

@OrcVole
Copy link

OrcVole commented May 5, 2023

https://huggingface.co/reeducator/vicuna-13b-free/discussions

It tries to reduce censorship

@fishscene
Copy link

Google PaLM looks interesting.
https://9to5google.com/2023/05/10/google-palm-2/

However, as far as I can tell, the framework is open source and might be available? But I can’t find any info on the training data. I’m also very new to AI and I might be looking in the wrong places for the data needed to be usable for Serge.

@rendel
Copy link

rendel commented May 11, 2023

MPT-7B: https://github.com/mosaicml/llm-foundry

@noproto
Copy link
Contributor

noproto commented May 25, 2023

I'd really like to see support for Guanaco models.
Edit: PR #334

@amiravni
Copy link

amiravni commented Jun 2, 2023

@Betanu701
Copy link

from my testing, it seems like all the Q4_0 and the Q8_0 work the best/fastest. The K varieties take too long with CPU only.
Honestly I think it is a toss up between Q4 and Q8. Both seem to run about the same within a margin of error.

@wuast94
Copy link

wuast94 commented Jun 18, 2023

@specked
Copy link

specked commented Jul 8, 2023

CodeGen2.5

@kagrith
Copy link

kagrith commented Jul 12, 2023

from my testing, it seems like all the Q4_0 and the Q8_0 work the best/fastest. The K varieties take too long with CPU only. Honestly I think it is a toss up between Q4 and Q8. Both seem to run about the same within a margin of error.

I get this odd error when trying to run Wizard.

llama.cpp: loading model from /usr/src/app/weights/Wizard-30B-Q4_1.bin
error loading model: unknown (magic, version) combination: 67676a74, 00000003; is this really a GGML file?
llama_init_from_file: failed to load model

Did you get this?
How did you resolve?

@tonyhardcode
Copy link

WizardCoder would be nice to have!

https://huggingface.co/WizardLM/WizardCoder-15B-V1.0

@laurentgoncalves
Copy link

Very interested to have the just released code llama : https://ai.meta.com/blog/code-llama-large-language-model-coding/

@gaby
Copy link
Member

gaby commented Nov 17, 2023

Fixed via #866

@gaby gaby closed this as completed Nov 17, 2023
@serge-chat serge-chat locked as resolved and limited conversation to collaborators Nov 17, 2023
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests