From f4a77bb73e008b2496d50c744aed7ce4c5d51b0f Mon Sep 17 00:00:00 2001 From: Xiaoyu Date: Thu, 30 Jan 2025 04:00:07 +0000 Subject: [PATCH] Fix nit --- .../configure-workflows/model-opt-and-transform/pytorch.md | 2 +- examples/llama2/llama2_lmeval.json | 2 +- examples/llama2/llama2_qlora.json | 2 +- olive/cli/finetune.py | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/source/how-to/configure-workflows/model-opt-and-transform/pytorch.md b/docs/source/how-to/configure-workflows/model-opt-and-transform/pytorch.md index e4431d144..c0652a64d 100644 --- a/docs/source/how-to/configure-workflows/model-opt-and-transform/pytorch.md +++ b/docs/source/how-to/configure-workflows/model-opt-and-transform/pytorch.md @@ -14,7 +14,7 @@ This pass only supports HfModels. Please refer to [LoRA](lora) for more details ```json { "type": "LoRA", - "lora_alpha": 16, + "alpha": 16, "train_data_config": // ..., "training_args": { "learning_rate": 0.0002, diff --git a/examples/llama2/llama2_lmeval.json b/examples/llama2/llama2_lmeval.json index b6630d25d..8328b25d9 100644 --- a/examples/llama2/llama2_lmeval.json +++ b/examples/llama2/llama2_lmeval.json @@ -42,7 +42,7 @@ "max_steps": 150, "logging_steps": 50.0 }, - "lora_alpha": 16, + "alpha": 16, "eval_data_config": "eval_data" } }, diff --git a/examples/llama2/llama2_qlora.json b/examples/llama2/llama2_qlora.json index ff0b97343..d5a0ae217 100644 --- a/examples/llama2/llama2_qlora.json +++ b/examples/llama2/llama2_qlora.json @@ -34,7 +34,7 @@ "logging_steps": 50.0 }, "r": 64, - "lora_alpha": 16, + "alpha": 16, "eval_data_config": "eval_data" }, "c": { diff --git a/olive/cli/finetune.py b/olive/cli/finetune.py index d83847150..70505f39b 100644 --- a/olive/cli/finetune.py +++ b/olive/cli/finetune.py @@ -108,7 +108,7 @@ def _get_run_config(self, tempdir: str) -> Dict: ((*finetune_key, "torch_dtype"), self.args.torch_dtype), ((*finetune_key, "training_args"), self.parse_training_args()), ((*finetune_key, "r"), self.args.lora_r), - ((*finetune_key, "lora_alpha"), self.args.lora_alpha), + ((*finetune_key, "alpha"), self.args.lora_alpha), ("output_dir", self.args.output_path), ("log_severity_level", self.args.log_level), ]