Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit 6ee413d

Browse files
committed
better way to handler jsoncpp
1 parent 6ceb477 commit 6ee413d

File tree

1 file changed

+13
-17
lines changed

1 file changed

+13
-17
lines changed

controllers/llamaCPP.cc

Lines changed: 13 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,8 @@ void llamaCPP::chatCompletion(
7777

7878
const auto &jsonBody = req->getJsonObject();
7979
std::string formatted_output =
80-
"Below is a conversation between an AI system named ASSISTANT and USER\n";
80+
"Below is a conversation between an AI system named " + ai_prompt +
81+
" and " + user_prompt + "\n";
8182

8283
json data;
8384
json stopWords;
@@ -106,6 +107,7 @@ void llamaCPP::chatCompletion(
106107
role = input_role;
107108
}
108109

110+
LOG_INFO << "VALUE OF ROLE BEING USED:" << role;
109111
std::string content = message["content"].asString();
110112
formatted_output += role + content + "\n";
111113
}
@@ -212,25 +214,19 @@ void llamaCPP::loadModel(
212214
LOG_INFO << "Drogon thread is:" << drogon_thread;
213215
if (jsonBody) {
214216
params.model = (*jsonBody)["llama_model_path"].asString();
215-
params.n_gpu_layers = (*jsonBody)["ngl"].asInt();
216-
params.n_ctx = (*jsonBody)["ctx_len"].asInt();
217-
params.embedding = (*jsonBody)["embedding"].asBool();
217+
params.n_gpu_layers = (*jsonBody).get("ngl", 100).asInt();
218+
params.n_ctx = (*jsonBody).get("ctx_len", 2048).asInt();
219+
params.embedding = (*jsonBody).get("embedding", true).asBool();
218220
// Check if n_parallel exists in jsonBody, if not, set to drogon_thread
219-
if ((*jsonBody).isMember("n_parallel")) {
220-
params.n_parallel = (*jsonBody)["n_parallel"].asInt();
221-
} else {
222-
params.n_parallel = drogon_thread;
223-
}
221+
222+
params.n_parallel = (*jsonBody).get("n_parallel", drogon_thread).asInt();
224223

225224
params.cont_batching = (*jsonBody)["cont_batching"].asBool();
226-
227-
// Set up prompt
228-
user_prompt = (*jsonBody)["user_prompt"].asString();
229-
ai_prompt = (*jsonBody)["ai_prompt"].asString();
230-
system_prompt = (*jsonBody)["system_prompt"].asString();
231-
232-
// params.n_threads = (*jsonBody)["n_threads"].asInt();
233-
// params.n_threads_batch = params.n_threads;
225+
226+
this->user_prompt = (*jsonBody).get("user_prompt", "USER: ").asString();
227+
this->ai_prompt = (*jsonBody).get("ai_prompt", "ASSISTANT: ").asString();
228+
this->system_prompt =
229+
(*jsonBody).get("system_prompt", "ASSISTANT's RULE: ").asString();
234230
}
235231
#ifdef GGML_USE_CUBLAS
236232
LOG_INFO << "Setting up GGML CUBLAS PARAMS";

0 commit comments

Comments
 (0)