mirror of
https://github.com/ollama/ollama.git
synced 2025-09-27 19:57:05 +02:00
do not cache prompt (#2018)
- prompt cache causes inferance to hang after some time
This commit is contained in:
@@ -181,7 +181,6 @@ func (llm *dynExtServer) Predict(ctx context.Context, predict PredictOpts, fn fu
|
|||||||
"seed": predict.Options.Seed,
|
"seed": predict.Options.Seed,
|
||||||
"stop": predict.Options.Stop,
|
"stop": predict.Options.Stop,
|
||||||
"image_data": imageData,
|
"image_data": imageData,
|
||||||
"cache_prompt": true,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if predict.Format == "json" {
|
if predict.Format == "json" {
|
||||||
|
Reference in New Issue
Block a user