We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Describe the bug
diff --git a/swift/llm/export/quant.py b/swift/llm/export/quant.py index 4e598c03..f3e17cde 100644 --- a/swift/llm/export/quant.py +++ b/swift/llm/export/quant.py @@ -210,6 +210,7 @@ class QuantEngine(ProcessorMixin): bits=args.quant_bits, dataset=','.join(args.dataset), batch_size=args.quant_batch_size, + group_size=args.group_size, block_name_to_quantize=self.get_block_name_to_quantize(self.model, args.model_type)) gptq_quantizer.serialization_keys.append('block_name_to_quantize') logger.info('Start quantizing the model...')
Your hardware and system info Write your system info like CUDA version/system/GPU/torch version here(在这里给出硬件信息和系统信息,如CUDA版本,系统,GPU型号和torch版本等)
Additional context Add any other context about the problem here(在这里补充其他信息)
The text was updated successfully, but these errors were encountered:
Sorry, I didn't understand what you meant.
Sorry, something went wrong.
Successfully merging a pull request may close this issue.
Describe the bug
Your hardware and system info
Write your system info like CUDA version/system/GPU/torch version here(在这里给出硬件信息和系统信息,如CUDA版本,系统,GPU型号和torch版本等)
Additional context
Add any other context about the problem here(在这里补充其他信息)
The text was updated successfully, but these errors were encountered: