Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the reference results are blank with deepseek model and our generate example code #12696

Open
K-Alex13 opened this issue Jan 10, 2025 · 4 comments
Assignees

Comments

@K-Alex13
Copy link

Following the guild of https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/example/GPU/HuggingFace/LLM/deepseek
and change the model to deepseek-ai/deepseek-coder-1.3b-instruct
following is the results
img_v3_02id_66a4783d-b64d-4bd4-80c2-edee438dbc1g
please help me

@K-Alex13
Copy link
Author

img_v3_02id_f2123c93-47d6-4c35-a542-9f91964032bg
even try to use original 6.7B model, a new problem come up

@Oscilloscope98
Copy link
Contributor

Hi @K-Alex13,

We are reproducing this issue, and will let you know for any updates :)

@Oscilloscope98
Copy link
Contributor

Hi @K-Alex13,

We have reproduced this issue and currently fixing it. We will update here for any progress :)

@Oscilloscope98
Copy link
Contributor

Oscilloscope98 commented Jan 16, 2025

Hi @K-Alex13,

We have fixed this blank output issue for deepseek-coder models, you could upgrade to ipex-llm>=2.2.0b20250115 and have a try again with the updated deepseek GPU example.

For your second issue regarding deepseek-ai/deepseek-coder-6.7b-instruct, this is because your memory is insufficient to load the original 6.7b model for further IPEX-LLM low-bit optimizations. You can use a machine with sufficient memory to save the IPEX-LLM low-bit model, then load the low-bit model onto the current machine for inference. You could refer to our Save-Load example for more information.

If no machine with larger memory is available, you could also try increasing your virtual memory to load the original mode.

Please let us know for any further problems :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants