Skip to content

Commit 32651d5

Browse files
Swati Allabadiochougul
Swati Allabadi
authored andcommitted
Using torch_qaic gradScaler and making lora_dropout=0.05 (#320)
1. In case of finetuning on qaic, torch_qaic gradScaler will be used 2. Moving back to lora_dropout = 0.05 on ML Framework team's ask. Signed-off-by: Swati Allabadi <[email protected]> Co-authored-by: Swati Allabadi <[email protected]>
1 parent 054a696 commit 32651d5

File tree

2 files changed

+6
-3
lines changed

2 files changed

+6
-3
lines changed

QEfficient/finetune/configs/peft_config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ class lora_config:
1919
target_modules: List[str] = field(default_factory=lambda: ["q_proj", "v_proj"])
2020
bias = "none"
2121
task_type: str = "CAUSAL_LM"
22-
lora_dropout: float = 0.0
22+
lora_dropout: float = 0.05
2323
inference_mode: bool = False # should be False for finetuning
2424

2525

QEfficient/finetune/utils/train_utils.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@
2424
import torch_qaic.debug as qaic_debug # noqa: F401
2525
import torch_qaic.profile as qaic_profile # noqa: F401
2626
import torch_qaic.utils as qaic_utils # noqa: F401
27+
from torch.qaic.amp import GradScaler as QAicGradScaler
2728
except ImportError as e:
2829
print(f"Warning: {e}. Moving ahead without these qaic modules.")
2930

@@ -60,7 +61,6 @@ def train(
6061
6162
Returns: results dictionary containing average training and validation perplexity and loss
6263
"""
63-
6464
train_prep = []
6565
train_loss = []
6666
val_prep = []
@@ -92,7 +92,10 @@ def train(
9292
tensorboard_updates = SummaryWriter()
9393

9494
if train_config.grad_scaler:
95-
scaler = GradScaler()
95+
if device.startswith("qaic"):
96+
scaler = QAicGradScaler()
97+
else:
98+
scaler = GradScaler()
9699

97100
loss_0_counter = torch.tensor([0]).to(device)
98101

0 commit comments

Comments
 (0)