You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I keep following the new code. I have tried to evaluate both the real quantized model and pseudo quantization using Llama-3-8B-Instruct model (w4-g128), based on wikitext-2-raw-v1(test). But the PPL result is 8.5. (GPU:A100, test data is downloaded from https://huggingface.co/datasets/Salesforce/wikitext/tree/main/wikitext-2-raw-v1)
I want to confirm if my result is correct? And why are the results different?
The text was updated successfully, but these errors were encountered:
I keep following the new code. I have tried to evaluate both the real quantized model and pseudo quantization using Llama-3-8B-Instruct model (w4-g128), based on wikitext-2-raw-v1(test). But the PPL result is 8.5. (GPU:A100, test data is downloaded from https://huggingface.co/datasets/Salesforce/wikitext/tree/main/wikitext-2-raw-v1)
I want to confirm if my result is correct? And why are the results different?
The text was updated successfully, but these errors were encountered: