From 294b6f5a220e8678c2b08fd2ab783a99e25c5215 Mon Sep 17 00:00:00 2001 From: frob Date: Tue, 21 Jan 2025 18:28:59 +0100 Subject: [PATCH] docs: remove tfs_z option from documentation (#8515) --- docs/modelfile.md | 1 - parser/parser_test.go | 1 - 2 files changed, 2 deletions(-) diff --git a/docs/modelfile.md b/docs/modelfile.md index b1c4e8a3c..5fcde7ed1 100644 --- a/docs/modelfile.md +++ b/docs/modelfile.md @@ -155,7 +155,6 @@ PARAMETER | temperature | The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8) | float | temperature 0.7 | | seed | Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt. (Default: 0) | int | seed 42 | | stop | Sets the stop sequences to use. When this pattern is encountered the LLM will stop generating text and return. Multiple stop patterns may be set by specifying multiple separate `stop` parameters in a modelfile. | string | stop "AI assistant:" | -| tfs_z | Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1) | float | tfs_z 1 | | num_predict | Maximum number of tokens to predict when generating text. (Default: -1, infinite generation) | int | num_predict 42 | | top_k | Reduces the probability of generating nonsense. A higher value (e.g. 100) will give more diverse answers, while a lower value (e.g. 10) will be more conservative. (Default: 40) | int | top_k 40 | | top_p | Works together with top-k. A higher value (e.g., 0.95) will lead to more diverse text, while a lower value (e.g., 0.5) will generate more focused and conservative text. (Default: 0.9) | float | top_p 0.9 | diff --git a/parser/parser_test.go b/parser/parser_test.go index eb52ee679..94f338ed2 100644 --- a/parser/parser_test.go +++ b/parser/parser_test.go @@ -490,7 +490,6 @@ func TestParseFileParameters(t *testing.T) { "top_k 1": {"top_k", "1"}, "top_p 1.0": {"top_p", "1.0"}, "min_p 0.05": {"min_p", "0.05"}, - "tfs_z 1.0": {"tfs_z", "1.0"}, "typical_p 1.0": {"typical_p", "1.0"}, "repeat_last_n 1": {"repeat_last_n", "1"}, "temperature 1.0": {"temperature", "1.0"},