diff --git a/pyproject.toml b/pyproject.toml index 082fc899..d6049724 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -12,7 +12,7 @@ build-backend = "poetry.masonry.api" [tool.poetry] name = "together" -version = "1.3.8" +version = "1.3.9" authors = [ "Together AI " ] diff --git a/src/together/cli/api/finetune.py b/src/together/cli/api/finetune.py index ff1a09b3..0eb2b00c 100644 --- a/src/together/cli/api/finetune.py +++ b/src/together/cli/api/finetune.py @@ -92,7 +92,7 @@ def fine_tuning(ctx: click.Context) -> None: @click.option( "--lora/--no-lora", type=bool, - default=False, + default=True, help="Whether to use LoRA adapters for fine-tuning", ) @click.option("--lora-r", type=int, default=8, help="LoRA adapters' rank") diff --git a/src/together/resources/finetune.py b/src/together/resources/finetune.py index 368a2506..b58cdae2 100644 --- a/src/together/resources/finetune.py +++ b/src/together/resources/finetune.py @@ -149,7 +149,7 @@ def create( warmup_ratio: float = 0.0, max_grad_norm: float = 1.0, weight_decay: float = 0.0, - lora: bool = False, + lora: bool = True, lora_r: int | None = None, lora_dropout: float | None = 0, lora_alpha: float | None = None, @@ -490,7 +490,7 @@ async def create( warmup_ratio: float = 0.0, max_grad_norm: float = 1.0, weight_decay: float = 0.0, - lora: bool = False, + lora: bool = True, lora_r: int | None = None, lora_dropout: float | None = 0, lora_alpha: float | None = None,