@@ -77,7 +77,8 @@ void llamaCPP::chatCompletion(
7777
7878 const auto &jsonBody = req->getJsonObject ();
7979 std::string formatted_output =
80- " Below is a conversation between an AI system named ASSISTANT and USER\n " ;
80+ " Below is a conversation between an AI system named " + ai_prompt +
81+ " and " + user_prompt + " \n " ;
8182
8283 json data;
8384 json stopWords;
@@ -106,6 +107,7 @@ void llamaCPP::chatCompletion(
106107 role = input_role;
107108 }
108109
110+ LOG_INFO << " VALUE OF ROLE BEING USED:" << role;
109111 std::string content = message[" content" ].asString ();
110112 formatted_output += role + content + " \n " ;
111113 }
@@ -212,25 +214,19 @@ void llamaCPP::loadModel(
212214 LOG_INFO << " Drogon thread is:" << drogon_thread;
213215 if (jsonBody) {
214216 params.model = (*jsonBody)[" llama_model_path" ].asString ();
215- params.n_gpu_layers = (*jsonBody)[ " ngl" ] .asInt ();
216- params.n_ctx = (*jsonBody)[ " ctx_len" ] .asInt ();
217- params.embedding = (*jsonBody)[ " embedding" ] .asBool ();
217+ params.n_gpu_layers = (*jsonBody). get ( " ngl" , 100 ) .asInt ();
218+ params.n_ctx = (*jsonBody). get ( " ctx_len" , 2048 ) .asInt ();
219+ params.embedding = (*jsonBody). get ( " embedding" , true ) .asBool ();
218220 // Check if n_parallel exists in jsonBody, if not, set to drogon_thread
219- if ((*jsonBody).isMember (" n_parallel" )) {
220- params.n_parallel = (*jsonBody)[" n_parallel" ].asInt ();
221- } else {
222- params.n_parallel = drogon_thread;
223- }
221+
222+ params.n_parallel = (*jsonBody).get (" n_parallel" , drogon_thread).asInt ();
224223
225224 params.cont_batching = (*jsonBody)[" cont_batching" ].asBool ();
226-
227- // Set up prompt
228- user_prompt = (*jsonBody)[" user_prompt" ].asString ();
229- ai_prompt = (*jsonBody)[" ai_prompt" ].asString ();
230- system_prompt = (*jsonBody)[" system_prompt" ].asString ();
231-
232- // params.n_threads = (*jsonBody)["n_threads"].asInt();
233- // params.n_threads_batch = params.n_threads;
225+
226+ this ->user_prompt = (*jsonBody).get (" user_prompt" , " USER: " ).asString ();
227+ this ->ai_prompt = (*jsonBody).get (" ai_prompt" , " ASSISTANT: " ).asString ();
228+ this ->system_prompt =
229+ (*jsonBody).get (" system_prompt" , " ASSISTANT's RULE: " ).asString ();
234230 }
235231#ifdef GGML_USE_CUBLAS
236232 LOG_INFO << " Setting up GGML CUBLAS PARAMS" ;
0 commit comments