mirror of
https://github.com/ollama/ollama.git
synced 2025-03-29 11:11:47 +01:00
fix sched to wait for the runner to terminate to ensure following vram check will be more accurate
This commit is contained in:
parent
a3e60d9058
commit
948114e3e3
@ -899,7 +899,10 @@ func (s *llmServer) Detokenize(ctx context.Context, tokens []int) (string, error
|
||||
func (s *llmServer) Close() error {
|
||||
if s.cmd != nil {
|
||||
slog.Debug("stopping llama server")
|
||||
return s.cmd.Process.Kill()
|
||||
if err := s.cmd.Process.Kill(); err != nil {
|
||||
return err
|
||||
}
|
||||
return s.cmd.Wait()
|
||||
}
|
||||
|
||||
return nil
|
||||
|
Loading…
x
Reference in New Issue
Block a user