-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dependencies are wrong #19
Comments
Hi, what do you mean by "very bad results"? As also discussed in #7, pruning LLaMa seems to be more challenging than pruning e.g. OPT, possibly because it is more parameter efficient. I just ran |
Hello, I make evaluations on some standart LLM evaluation tasks, using "LLM Evaluation Harness": I get the following results for LLaMA:
Normally I would expect some performance drop yet for comparison here are the results for BLOOM-7B:
So probably there is something wrong with the implementation, as I mentioned my own implementation also get the same results, so I would like to compare it with your results. Can you please do the experiments with the latest version of |
Sorry I've closed the issue by accident, I would be glad if you can re-open so we can solve the issue. I also tried the experiment with multiple torch, python and transformers versions, if your results are better, I would expect that it works for a very spesific version of libraries for some reason. |
Hello, I have tried lots of different version combinations to make the LLaMA script work, it produces very bad results which is
also what I observed with my own implementation and some other implementation for SparseGPT LLaMA.
All 3 of these implementations produce exactly the same results, which is good it shows probably we are doing everything correctly,
but then the performance is incredibly poor for LLaMA, it performs even worse than BLOOM or OPT.
If your results are better can you please share the exact dependencies to repeat your experiments, because the transformers
library version you give in the README does not even have LLaMA tokenizer etc.
Thank you
The text was updated successfully, but these errors were encountered: