-
Notifications
You must be signed in to change notification settings - Fork 188
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reproduce the results of the code in the paper #30
Comments
Hi, To properly evaluate the quantized model on downstream tasks, you should use the lm-evaluation-harness framework. This framework is designed to be compatible with Hugging Face Transformers and provides a standardized way to run various benchmarks. |
Thank you Professor, I know the framework. But this framework is also needed to quantise the model after I get it and then do the inference task. |
Hi, it actually will be quantized when you try to inference it. If you check our code closely you will find it will be quantized during the inference. |
Dear Professor, I followed what you said about framing this lm-evaluation-harness , and tried to rewrite this code. import argparse import torch sys.path.append("./lm-evaluation-harness") try:
except ModuleNotFoundError: logging.getLogger("openai").setLevel(logging.WARNING) class MultiChoice:
def parse_args():
Returns a list containing all values of the source_list thatmatch at least one of the patternsdef pattern_match(patterns, source_list): def main():
if name == "main": |
How do you reproduce the results of the code in the paper in your code? According to your example, I get the following result when I execute "generate".
![image](https://private-user-images.githubusercontent.com/149936473/343762081-c22a1f33-e220-4ead-ae77-2d0b5feb5f9c.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzkzNDc1MjEsIm5iZiI6MTczOTM0NzIyMSwicGF0aCI6Ii8xNDk5MzY0NzMvMzQzNzYyMDgxLWMyMmExZjMzLWUyMjAtNGVhZC1hZTc3LTJkMGI1ZmViNWY5Yy5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjUwMjEyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI1MDIxMlQwODAwMjFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT03OWVkMDM5ZGZiZmU5NWE4N2M0MGQ3N2Y1ZGQyYjZhN2Q1ZTg4NjBkNDBiZDhiNmVhZDdhYTU2MTFkNWQwMWVhJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCJ9.nAHFZEYJ4DiVaPkjDTzGC2YudRDPWPQ_MynTvXhJ4ZM)
But I still can't feel the quantisation in the paper on the downstream task.
The text was updated successfully, but these errors were encountered: