mirror of
https://github.com/ollama/ollama.git
synced 2025-03-27 10:11:54 +01:00
The OLLAMA_MAX_VRAM env var was a temporary workaround for OOM scenarios. With Concurrency this was no longer wired up, and the simplistic value doesn't map to multi-GPU setups. Users can still set `num_gpu` to limit memory usage to avoid OOM if we get our predictions wrong.
Integration Tests
This directory contains integration tests to exercise Ollama end-to-end to verify behavior
By default, these tests are disabled so go test ./...
will exercise only unit tests. To run integration tests you must pass the integration tag. go test -tags=integration ./...
The integration tests have 2 modes of operating.
- By default, they will start the server on a random port, run the tests, and then shutdown the server.
- If
OLLAMA_TEST_EXISTING
is set to a non-empty string, the tests will run against an existing running server, which can be remote