Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question on the Experiment Results #2

Open
SJTU-EPIC-Lab opened this issue Jan 3, 2025 · 1 comment
Open

Question on the Experiment Results #2

SJTU-EPIC-Lab opened this issue Jan 3, 2025 · 1 comment

Comments

@SJTU-EPIC-Lab
Copy link

Thank you for the most valuable implementation!!!!!

Our lab is also interested in this work and we have tried your implementation. However, we find a very significant accuracy drop when LazyLLM is applied (narrative 17.89->13,24). How about your results? Do you get the reported results in the papers? We are quite glad for your response.

Best Regards

@Adam-Mazur
Copy link
Owner

Hello,
I'm glad my code came in came in useful. Unfortunately, I haven't done any extensive tests of those models, since this is only a hobby project. However, if you are getting significant accuracy drops it's likely due to suboptimal pruning rates' configuration. If I recall correctly, the authors of the original paper mentioned that they achived the best results with pruning rates decreasing from the first layer to the last layer. In the example code I provided the pruning rates are fixed at 0.1 for every layer, which could explain the accuracy drop.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants