@@ -33,13 +33,13 @@ def _add_create(parser: argparse._SubParsersAction[argparse.ArgumentParser]) ->
33
33
required = True ,
34
34
type = str ,
35
35
)
36
- create_finetune_parser .add_argument (
37
- "--validation-file" ,
38
- "-v" ,
39
- default = None ,
40
- help = "The ID of an uploaded file that contains validation data." ,
41
- type = str ,
42
- )
36
+ # create_finetune_parser.add_argument(
37
+ # "--validation-file",
38
+ # "-v",
39
+ # default=None,
40
+ # help="The ID of an uploaded file that contains validation data.",
41
+ # type=str,
42
+ # )
43
43
create_finetune_parser .add_argument (
44
44
"--model" ,
45
45
"-m" ,
@@ -57,53 +57,57 @@ def _add_create(parser: argparse._SubParsersAction[argparse.ArgumentParser]) ->
57
57
create_finetune_parser .add_argument (
58
58
"--batch-size" ,
59
59
"-b" ,
60
- default = None ,
60
+ default = 32 ,
61
61
help = "The batch size to use for training." ,
62
62
type = int ,
63
63
)
64
64
create_finetune_parser .add_argument (
65
- "--learning-rate-multiplier " ,
66
- "-lrm " ,
67
- default = None ,
65
+ "--learning-rate" ,
66
+ "-lr " ,
67
+ default = 0.00001 ,
68
68
help = "The learning rate multiplier to use for training." ,
69
69
type = float ,
70
70
)
71
- create_finetune_parser .add_argument (
72
- "--prompt-loss-weight" ,
73
- "-plw" ,
74
- default = 0.01 ,
75
- help = "The weight to use for loss on the prompt tokens." ,
76
- type = float ,
77
- )
78
- create_finetune_parser .add_argument (
79
- "--compute-classification-metrics" ,
80
- "-ccm" ,
81
- default = False ,
82
- action = "store_true" ,
83
- help = "Calculate classification-specific metrics using the validation set." ,
84
- )
85
- create_finetune_parser .add_argument (
86
- "--classification-n-classes" ,
87
- "-cnc" ,
88
- default = None ,
89
- help = "The number of classes in a classification task." ,
90
- type = int ,
91
- )
92
- create_finetune_parser .add_argument (
93
- "--classification-positive-class" ,
94
- "-cpc" ,
95
- default = None ,
96
- help = "The positive class in binary classification." ,
97
- type = str ,
98
- )
99
- create_finetune_parser .add_argument (
100
- "--classification-betas" ,
101
- "-cb" ,
102
- default = None ,
103
- help = "Calculate F-beta scores at the specified beta values." ,
104
- nargs = "+" ,
105
- type = float ,
106
- )
71
+ # create_finetune_parser.add_argument(
72
+ # "--warmup-steps",
73
+ # "-ws",
74
+ # default=0,
75
+ # help="Warmup steps",
76
+ # type=int,
77
+ # )
78
+ # create_finetune_parser.add_argument(
79
+ # "--train-warmup-steps",
80
+ # "-tws",
81
+ # default=0,
82
+ # help="Train warmup steps",
83
+ # type=int,
84
+ # )
85
+ # create_finetune_parser.add_argument(
86
+ # "--sequence-length",
87
+ # "-sl",
88
+ # default=2048,
89
+ # help="Max sequence length",
90
+ # type=int,
91
+ # )
92
+ # create_finetune_parser.add_argument(
93
+ # "--seed",
94
+ # default=42,
95
+ # help="Training seed",
96
+ # type=int,
97
+ # )
98
+ # create_finetune_parser.add_argument(
99
+ # "--fp32",
100
+ # help="Enable FP32 training. Defaults to false (FP16 training).",
101
+ # default=False,
102
+ # action="store_true",
103
+ # )
104
+ # create_finetune_parser.add_argument(
105
+ # "--checkpoint-steps",
106
+ # "-b",
107
+ # default=0,
108
+ # help="Number of steps between each checkpoint. Defaults to 0 = checkpoints per epoch.",
109
+ # type=int,
110
+ # )
107
111
create_finetune_parser .add_argument (
108
112
"--suffix" ,
109
113
"-s" ,
@@ -244,16 +248,17 @@ def _run_create(args: argparse.Namespace) -> None:
244
248
245
249
response = finetune .create_finetune (
246
250
training_file = args .training_file , # training file_id
247
- validation_file = args .validation_file , # validation file_id
251
+ # validation_file=args.validation_file, # validation file_id
248
252
model = args .model ,
249
253
n_epochs = args .n_epochs ,
250
254
batch_size = args .batch_size ,
251
- learning_rate_multiplier = args .learning_rate_multiplier ,
252
- prompt_loss_weight = args .prompt_loss_weight ,
253
- compute_classification_metrics = args .compute_classification_metrics ,
254
- classification_n_classes = args .classification_n_classes ,
255
- classification_positive_class = args .classification_positive_class ,
256
- classification_betas = args .classification_betas ,
255
+ learning_rate = args .learning_rate ,
256
+ # warmup_steps=args.warmup_steps,
257
+ # train_warmup_steps=args.train_warmup_steps,
258
+ # seq_length=args.sequence_length,
259
+ # seed=args.seed,
260
+ # fp16=not args.fp32,
261
+ # checkpoint_steps=args.checkpoint_steps,
257
262
suffix = args .suffix ,
258
263
)
259
264
0 commit comments