Skip to content
This repository has been archived by the owner on Aug 30, 2024. It is now read-only.

[GPTQ Enhance] Fix convert_quantized model.py for dtype #158

Merged
merged 1 commit into from
Mar 8, 2024

Conversation

zhentaoyu
Copy link
Contributor

Type of Change

bug fix

Description

detail description
Issues: xxx

Expected Behavior & Potential Risk

the expected behavior that triggered by this PR

How has this PR been tested?

how to reproduce the test (including hardware information)

Dependency Change?

any library dependency introduced or removed

@zhentaoyu zhentaoyu changed the title fix convert_quantized model bug [Neural Speed] Fix convert_quantized model bug Mar 7, 2024
@Zhenzhong1 Zhenzhong1 changed the title [Neural Speed] Fix convert_quantized model bug [GPTQ Enhence] Fix convert_quantized model bug Mar 7, 2024
Copy link
Contributor

@a32543254 a32543254 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@a32543254 a32543254 changed the title [GPTQ Enhence] Fix convert_quantized model bug [GPTQ Enhence] Fix convert_quantized model.py for dtype Mar 7, 2024
@zhentaoyu zhentaoyu changed the title [GPTQ Enhence] Fix convert_quantized model.py for dtype [GPTQ Enhance] Fix convert_quantized model.py for dtype Mar 7, 2024
@VincyZhang VincyZhang merged commit 37d01f3 into main Mar 8, 2024
11 checks passed
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants