diff --git a/README.md b/README.md index 1924fe5..a439f0a 100644 --- a/README.md +++ b/README.md @@ -88,6 +88,20 @@ This project is based on the [llama.cpp](https://github.com/ggerganov/llama.cpp)
usage: run_inference.py [-h] [-m MODEL] [-n N_PREDICT] -p PROMPT [-t THREADS] [-c CTX_SIZE] [-temp TEMPERATURE] [-cnv]
@@ -186,6 +195,7 @@ optional arguments:
-temp TEMPERATURE, --temperature TEMPERATURE
Temperature, a hyperparameter that controls the randomness of the generated text
-cnv, --conversation Whether to enable chat mode or not (for instruct models.)
+ (When this option is turned on, the prompt specified by -p will be used as the system prompt.)
### Benchmark