mirror of
https://github.com/deepseek-ai/DeepSeek-MoE.git
synced 2025-02-23 14:19:06 -05:00
fix: ValueError argument bug
Fix the duplicate argument `ValueError: You can't pass `load_in_4bit`or `load_in_8bit` as a kwarg when passing `quantization_config` argument at the same time.`
This commit is contained in:
parent
66edeee5a4
commit
d00af39123
@ -184,8 +184,6 @@ def build_model(model_args, training_args, checkpoint_dir):
|
|||||||
compute_dtype = (torch.bfloat16 if training_args.bf16 else torch.float16)
|
compute_dtype = (torch.bfloat16 if training_args.bf16 else torch.float16)
|
||||||
model = transformers.AutoModelForCausalLM.from_pretrained(
|
model = transformers.AutoModelForCausalLM.from_pretrained(
|
||||||
model_args.model_name_or_path,
|
model_args.model_name_or_path,
|
||||||
load_in_4bit=model_args.bits == 4,
|
|
||||||
load_in_8bit=model_args.bits == 8,
|
|
||||||
quantization_config=BitsAndBytesConfig(
|
quantization_config=BitsAndBytesConfig(
|
||||||
load_in_4bit=model_args.bits == 4,
|
load_in_4bit=model_args.bits == 4,
|
||||||
load_in_8bit=model_args.bits == 8,
|
load_in_8bit=model_args.bits == 8,
|
||||||
|
Loading…
Reference in New Issue
Block a user