@@ -364,12 +364,12 @@ void FlexFlow::top_level_task(Task const *task,
364364 num_layers_per_finetuning_step,
365365 temporal_sharing_frequency,
366366 run_warmup);
367- assert (peft_finetuning_enabled (ffconfig.peft_support_mode ) &&
368- " Cannot train LORA adapter if finetuning is not enabled" );
369- assert (!file_paths.dataset_file_path .empty () &&
370- " Cannot train LORA adapter if dataset path is empty" );
371- assert (!peft_model_name.empty () &&
372- " PEFT model name should not be left empty" );
367+ // assert(peft_finetuning_enabled(ffconfig.peft_support_mode) &&
368+ // "Cannot train LORA adapter if finetuning is not enabled");
369+ // assert(!file_paths.dataset_file_path.empty() &&
370+ // "Cannot train LORA adapter if dataset path is empty");
371+ // assert(!peft_model_name.empty() &&
372+ // "PEFT model name should not be left empty");
373373
374374 if (num_kv_cache_slots == -1 ) {
375375 num_kv_cache_slots = max_sequence_length * max_requests_per_batch;
@@ -520,17 +520,17 @@ void FlexFlow::top_level_task(Task const *task,
520520 rm->start_background_server (&model);
521521
522522 // Add PEFT adapter(s)
523- PEFTModelID *peft_model_id_finetuning =
524- model.register_peft_adapter (peft_config_finetuning);
525-
526- if (run_warmup) {
527- std::vector<Request> warmup_requests =
528- make_warmup_requests (10 , 1000 , peft_model_id_finetuning);
529- std::vector<GenerationResult> warmup_result =
530- model.generate (warmup_requests);
531- rm->set_inference_finished (false ); // reset inference finished flag
532- std::cout << " ----------warmup finished--------------" << std::endl;
533- }
523+ // PEFTModelID *peft_model_id_finetuning =
524+ // model.register_peft_adapter(peft_config_finetuning);
525+
526+ // if (run_warmup) {
527+ // std::vector<Request> warmup_requests =
528+ // make_warmup_requests(10, 1000, peft_model_id_finetuning);
529+ // std::vector<GenerationResult> warmup_result =
530+ // model.generate(warmup_requests);
531+ // rm->set_inference_finished(false); // reset inference finished flag
532+ // std::cout << "----------warmup finished--------------" << std::endl;
533+ // }
534534
535535 // Run workload
536536 {
@@ -544,23 +544,23 @@ void FlexFlow::top_level_task(Task const *task,
544544 }
545545
546546 // Add fine-tuning request
547- assert (!file_paths.dataset_file_path .empty () &&
548- " Dataset file path is required for fine-tuning." );
549- printf (" Finetuning request with dataset %s\n " ,
550- file_paths.dataset_file_path .c_str ());
551- Request fine_tuning_req;
552- fine_tuning_req.req_type = RequestType::REQ_FINETUNING;
553- fine_tuning_req.peft_model_id = *peft_model_id_finetuning;
554- fine_tuning_req.peft_finetuning_info .dataset_filepath =
555- file_paths.dataset_file_path ;
556- fine_tuning_req.peft_finetuning_info .max_samples = max_finetuning_samples;
557- fine_tuning_req.peft_finetuning_info .max_training_epochs =
558- max_training_epochs;
559- fine_tuning_req.peft_finetuning_info .gradient_accumulation_steps =
560- gradient_accumulation_steps;
561- fine_tuning_req.peft_finetuning_info .num_logging_steps = num_logging_steps;
547+ // assert(!file_paths.dataset_file_path.empty() &&
548+ // "Dataset file path is required for fine-tuning.");
549+ // printf("Finetuning request with dataset %s\n",
550+ // file_paths.dataset_file_path.c_str());
551+ // Request fine_tuning_req;
552+ // fine_tuning_req.req_type = RequestType::REQ_FINETUNING;
553+ // fine_tuning_req.peft_model_id = *peft_model_id_finetuning;
554+ // fine_tuning_req.peft_finetuning_info.dataset_filepath =
555+ // file_paths.dataset_file_path;
556+ // fine_tuning_req.peft_finetuning_info.max_samples = max_finetuning_samples;
557+ // fine_tuning_req.peft_finetuning_info.max_training_epochs =
558+ // max_training_epochs;
559+ // fine_tuning_req.peft_finetuning_info.gradient_accumulation_steps =
560+ // gradient_accumulation_steps;
561+ // fine_tuning_req.peft_finetuning_info.num_logging_steps = num_logging_steps;
562562 std::vector<Request> finetuning_requests;
563- finetuning_requests.push_back (fine_tuning_req);
563+ // finetuning_requests.push_back(fine_tuning_req);
564564
565565 std::cout << " ----------inference started--------------" << std::endl;
566566 std::vector<GenerationResult> result =
@@ -598,7 +598,7 @@ void FlexFlow::top_level_task(Task const *task,
598598 run_warmup ? 10 : 0 ); // num_warmup_requests
599599 }
600600
601- free (peft_model_id_finetuning);
601+ // free(peft_model_id_finetuning);
602602
603603 std::cout << " ----------inference finished--------------" << std::endl;
604604
0 commit comments