|
@ -188,8 +188,8 @@ class FinetuningArguments: |
|
|
lora_target: Optional[str] = field( |
|
|
lora_target: Optional[str] = field( |
|
|
default="q_proj,v_proj", |
|
|
default="q_proj,v_proj", |
|
|
metadata={"help": "Name(s) of target modules to apply LoRA. Use comma to separate multiple modules. \ |
|
|
metadata={"help": "Name(s) of target modules to apply LoRA. Use comma to separate multiple modules. \ |
|
|
LLaMA choices: [\"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\", \"mlp\"], \ |
|
|
LLaMA choices: [\"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\", \"up_proj\", \"down_proj\"], \ |
|
|
BLOOM choices: [\"query_key_value\", \"dense\", \"mlp\"]"} |
|
|
BLOOM choices: [\"query_key_value\", \"dense\", \"dense_\"]"} |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
def __post_init__(self): |
|
|
def __post_init__(self): |
|
|