From 8372c95103b52cecff22dc89151af8201499167d Mon Sep 17 00:00:00 2001 From: Galaxy1458 <55453380+Galaxy1458@users.noreply.github.com> Date: Thu, 30 May 2024 18:22:37 +0800 Subject: [PATCH 1/3] Update merge_lora_params.py --- llm/merge_lora_params.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llm/merge_lora_params.py b/llm/merge_lora_params.py index 50ae4a797f34..663b25cc5f62 100644 --- a/llm/merge_lora_params.py +++ b/llm/merge_lora_params.py @@ -125,7 +125,7 @@ def merge(): model = AutoModelForCausalLM.from_pretrained( lora_config.base_model_name_or_path, config=config, - low_cpu_mem_usage=True, + low_cpu_mem_usage=False, ) lora_config.merge_weights = True model = LoRAModel.from_pretrained(model=model, lora_path=args.lora_path, lora_config=lora_config) From 244bec6ca01eaebbc9967270d78a8a87ec09c3f9 Mon Sep 17 00:00:00 2001 From: Galaxy1458 <55453380+Galaxy1458@users.noreply.github.com> Date: Thu, 30 May 2024 20:16:38 +0800 Subject: [PATCH 2/3] Update merge_lora_params.py --- llm/merge_lora_params.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llm/merge_lora_params.py b/llm/merge_lora_params.py index 663b25cc5f62..ebfd889abbe9 100644 --- a/llm/merge_lora_params.py +++ b/llm/merge_lora_params.py @@ -110,7 +110,7 @@ def merge(): model = AutoModelForCausalLM.from_pretrained( lora_config.base_model_name_or_path, config=config, - low_cpu_mem_usage=True, + low_cpu_mem_usage=args.low_gpu_mem, ) model = LoRAModel.from_pretrained(model=model, lora_path=args.lora_path, lora_config=lora_config) model.eval() From 6c1584aeaf1a4c4f811d07ec6985c3ea99a757a1 Mon Sep 17 00:00:00 2001 From: Galaxy1458 <55453380+Galaxy1458@users.noreply.github.com> Date: Thu, 30 May 2024 20:54:18 +0800 Subject: [PATCH 3/3] Update merge_lora_params.py --- llm/merge_lora_params.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llm/merge_lora_params.py b/llm/merge_lora_params.py index ebfd889abbe9..065a2585ebc0 100644 --- a/llm/merge_lora_params.py +++ b/llm/merge_lora_params.py @@ -110,7 +110,7 @@ def merge(): model = AutoModelForCausalLM.from_pretrained( lora_config.base_model_name_or_path, config=config, - low_cpu_mem_usage=args.low_gpu_mem, + low_cpu_mem_usage=True, ) model = LoRAModel.from_pretrained(model=model, lora_path=args.lora_path, lora_config=lora_config) model.eval() @@ -125,7 +125,7 @@ def merge(): model = AutoModelForCausalLM.from_pretrained( lora_config.base_model_name_or_path, config=config, - low_cpu_mem_usage=False, + low_cpu_mem_usage=args.low_gpu_mem, ) lora_config.merge_weights = True model = LoRAModel.from_pretrained(model=model, lora_path=args.lora_path, lora_config=lora_config)