Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add gpt4all local models, including an embedding provider #454

Merged
merged 6 commits into from
Nov 9, 2023

Conversation

3coins
Copy link
Collaborator

@3coins 3coins commented Nov 8, 2023

Description

Adds new local models from gpt4all. Also adds an embedding provider for local embeddings generation with /learn. Also, fixes a small bug with the base chat handler, all exceptions thrown are causing an error because the method is not awaited currently.

Note

Could not add the replit-code-v1_5-3b-q4_0.gguf model, which crashes the server and throws this error when chat is used. If you know how to fix this, please contribute or leave a suggestion to fix in this PR.

GGML_ASSERT: /Users/distiller/project/gpt4all-backend/llama.cpp-mainline/ggml-metal.m:1178: false && "only power-of-two n_head implemented"

@3coins 3coins added bug Something isn't working enhancement New feature or request labels Nov 8, 2023
@3coins 3coins marked this pull request as ready for review November 8, 2023 23:33
docs/source/users/index.md Outdated Show resolved Hide resolved
docs/source/users/index.md Show resolved Hide resolved
@3coins 3coins self-assigned this Nov 9, 2023
@krassowski
Copy link
Member

Regarding model size discussion just as FYI I wanted to highlight that there is a machine-readable JSON from which the download size, ram required, description, and prompt template for each model can be extracted as discussed in: #413

@3coins
Copy link
Collaborator Author

3coins commented Nov 9, 2023

@krassowski
Thanks for pointing to the model catalog. I agree, we could use some of the info from the machine-readable JSON, however, properties like prompt templates have to be formatted to work with LangChain and might need some work than directly importing them.

@3coins 3coins merged commit 56b189c into jupyterlab:main Nov 9, 2023
6 checks passed
@3coins
Copy link
Collaborator Author

3coins commented Nov 10, 2023

@meeseeksdev please backport to 1.x

meeseeksmachine pushed a commit to meeseeksmachine/jupyter-ai that referenced this pull request Nov 10, 2023
3coins added a commit that referenced this pull request Nov 10, 2023
dbelgrod pushed a commit to dbelgrod/jupyter-ai that referenced this pull request Jun 10, 2024
…#454)

* Added models for gpt4all.

* Removed replit model.

* Updated docs for gpt4all.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Updated docs.

* A fix to stop download of embeddings model.

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Marchlak pushed a commit to Marchlak/jupyter-ai that referenced this pull request Oct 28, 2024
…#454)

* Added models for gpt4all.

* Removed replit model.

* Updated docs for gpt4all.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Updated docs.

* A fix to stop download of embeddings model.

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants