Skip to content

Commit 0170f38

Browse files
committed
test fix
1 parent f8baf32 commit 0170f38

File tree

2 files changed

+8
-1
lines changed

2 files changed

+8
-1
lines changed

src/axolotl/loaders/model.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -556,8 +556,9 @@ def _set_quantization_config(self):
556556
self.cfg.adapter in ["qlora", "qalora"]
557557
and self.model_kwargs["load_in_4bit"]
558558
):
559+
quantization_config = getattr(self.model_config, "quantization_config", {})
559560
self.model_kwargs["quantization_config"] = BitsAndBytesConfig(
560-
**self.model_config.quantization_config
561+
**quantization_config
561562
)
562563
elif self.cfg.adapter == "qlora" and self.model_kwargs.get(
563564
"load_in_4bit", False

src/axolotl/utils/schemas/config.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -541,6 +541,12 @@ class AxolotlInputConfig(
541541
"description": "Whether to use flash-attention rms norm implementation - advanced use only"
542542
},
543543
)
544+
flash_attn_fuse_qkv: bool | None = Field(
545+
default=None,
546+
json_schema_extra={
547+
"description": "Whether to fuse QKV projection into a single operation"
548+
},
549+
)
544550
flash_attn_fuse_mlp: bool | None = Field(
545551
default=None,
546552
json_schema_extra={

0 commit comments

Comments
 (0)