Yashwanth A
291def6adb
server: increase timeout in stall detection from 5s to 30s ( #8831 )
...
In some cases, downloads slow due to disk i/o or other factors,
causing the download to restart a part. This causes the download
to "reverse" in percent completion. By increasing the timeout to 30s,
this should happen less frequently.
v0.5.8-rc7
2025-02-05 10:00:26 -08:00
Jeffrey Morgan
cd3fbf1c49
llama: use dynamic backend loading for mllama and clip ( #8835 )
2025-02-05 09:46:56 -08:00
Jeffrey Morgan
c852b8e021
server: always print upload/download part info ( #8832 )
2025-02-04 19:30:49 -08:00
William
d8932c55e7
server: fix out of bounds exception on model download ( #8746 )
2025-02-04 18:52:47 -08:00
Michael Yang
63f0269f7f
ci: split docker build by platform
...
this improves build reliability and concurrency
v0.5.8-rc6
2025-02-04 17:04:27 -08:00
Jeffrey Morgan
4759ecae19
ml/backend/ggml: fix library loading on macOS amd64 ( #8827 )
2025-02-04 15:05:39 -08:00
Michael Yang
65b7ecac7b
fix extra quote
v0.5.8-rc5
2025-02-04 08:35:30 -08:00
Michael Yang
f9d2d89135
fix linux archive
v0.5.8-rc4
2025-02-03 16:12:33 -08:00
Michael Yang
669dc31cf3
fix build
v0.5.8-rc3
2025-02-03 15:10:51 -08:00
Tilman Griesel
d4d338c224
readme: add Chipper to community integrations ( #8803 )
2025-02-03 14:18:19 -08:00
Melroy van den Berg
bfdeffc375
docs: use OLLAMA_VERSION=0.5.7 for install version override ( #8802 )
2025-02-03 13:54:08 -08:00
Michael Yang
e806184023
fix release workflow
v0.5.8-rc2
2025-02-03 13:19:57 -08:00
Jeffrey Morgan
50566113ac
llm: do not error if LibOllamaPath does not exist ( #8801 )
2025-02-03 12:27:48 -08:00
Davide Bertoni
ad22ace439
docs: add missing json and shell code blocks in api.md ( #8766 )
2025-02-02 13:12:55 -08:00
Anıl Kaynar
f4321a421c
readme: add MinimalNextOllamaChat to community integrations ( #8767 )
2025-02-02 12:56:10 -08:00
Michael Yang
475333d533
fix docker build-args
...
env context is not accessible from job.*.strategy. since it's in the
environment, just tell docker to use the environment variable[1]
[1]: https://docs.docker.com/reference/cli/docker/buildx/build/#build-arg
v0.5.8-rc1
2025-01-31 14:56:02 -08:00
Michael Yang
39fd89308c
build: set CFLAGS=-O3 specifically for cpu.go
v0.5.8-rc0
2025-01-31 10:25:39 -08:00
Michael Yang
548a9f56a6
Revert "cgo: use O3"
...
This reverts commit bea1f1fac6b6b51bb3b8a666789c518b7aaa8b94.
2025-01-31 10:25:39 -08:00
Michael Yang
3f0cb36bdb
build: set goflags in linux release
2025-01-30 13:07:32 -08:00
Michael Yang
bea1f1fac6
cgo: use O3
2025-01-30 12:21:50 -08:00
Jeffrey Morgan
5d75d837ef
discover: fix default LibOllamaPath value ( #8702 )
2025-01-30 12:21:38 -08:00
Parth Sareen
711648c9bb
docs: update api.md with streaming with tools is enabled ( #8676 )
2025-01-29 15:14:30 -08:00
Michael Yang
dcfb7a105c
next build ( #8539 )
...
* add build to .dockerignore
* test: only build one arch
* add build to .gitignore
* fix ccache path
* filter amdgpu targets
* only filter if autodetecting
* Don't clobber gpu list for default runner
This ensures the GPU specific environment variables are set properly
* explicitly set CXX compiler for HIP
* Update build_windows.ps1
This isn't complete, but is close. Dependencies are missing, and it only builds the "default" preset.
* build: add ollama subdir
* add .git to .dockerignore
* docs: update development.md
* update build_darwin.sh
* remove unused scripts
* llm: add cwd and build/lib/ollama to library paths
* default DYLD_LIBRARY_PATH to LD_LIBRARY_PATH in runner on macOS
* add additional cmake output vars for msvc
* interim edits to make server detection logic work with dll directories like lib/ollama/cuda_v12
* remove unncessary filepath.Dir, cleanup
* add hardware-specific directory to path
* use absolute server path
* build: linux arm
* cmake install targets
* remove unused files
* ml: visit each library path once
* build: skip cpu variants on arm
* build: install cpu targets
* build: fix workflow
* shorter names
* fix rocblas install
* docs: clean up development.md
* consistent build dir removal in development.md
* silence -Wimplicit-function-declaration build warnings in ggml-cpu
* update readme
* update development readme
* llm: update library lookup logic now that there is one runner (#8587 )
* tweak development.md
* update docs
* add windows cuda/rocm tests
---------
Co-authored-by: jmorganca <jmorganca@gmail.com>
Co-authored-by: Daniel Hiltgen <daniel@ollama.com>
2025-01-29 15:03:38 -08:00
Xiaofu Huang
2ef3c803a1
readme: add AI Toolkit for VSCode to community integrations ( #8604 )
2025-01-27 00:36:23 -08:00
Matěj Štágl
453e4d090b
readme: add LlmTornado to community integrations ( #8551 )
2025-01-25 01:04:07 -08:00
Daniel Jalkut
ca2f9843c8
docs: remove reference to the deleted examples folder ( #8524 )
2025-01-22 22:52:15 -08:00
frob
294b6f5a22
docs: remove tfs_z option from documentation ( #8515 )
2025-01-21 09:28:59 -08:00
EndoTheDev
7bb356c680
docs: update suspend header in gpu.md ( #8487 )
2025-01-19 18:45:35 -08:00
Jannik Maierhöfer
021817e59a
readme: add link to Langfuse ( #8455 )
2025-01-16 22:41:12 -08:00
Patrick Devine
a420a453b4
fix default modelfile for create ( #8452 )
v0.5.7
2025-01-16 01:14:04 -08:00
Jeffrey Morgan
42cf4db601
parser: fix parsing Modelfiles with multiple FROM commands ( #8449 )
2025-01-16 00:14:04 -08:00
Josh
93a8daf285
convert: import support for command-r models from safetensors ( #6063 )
...
---------
Co-authored-by: Patrick Devine <patrick@infrahq.com>
2025-01-15 16:31:22 -08:00
Gloryjaw
a041b4df7c
docs: fix path to examples ( #8438 )
2025-01-15 11:49:12 -08:00
Patrick Devine
2539f2dbf9
Fix absolute path names + gguf detection ( #8428 )
v0.5.6
2025-01-14 19:01:24 -08:00
Jeffrey Morgan
61676fb506
llama: move grammar tests to llama_test.go ( #8411 )
2025-01-14 12:55:45 -08:00
Bruce MacDonald
f6f3713001
convert: qwen2 from safetensors ( #8408 )
...
Add native support for converting Qwen2 family models (including Qwen2.5)
from safetensors to gguf format so we can run it.
2025-01-14 10:34:37 -08:00
Steve Berdy
a30f347201
readme: add LangChain for .NET to community integrations ( #8352 )
2025-01-14 09:37:35 -08:00
Jeffrey Morgan
74ea4fb604
remove .prettierrc.json ( #8413 )
2025-01-14 09:30:34 -08:00
Jeffrey Morgan
6982e9cc96
readme: remove link to missing page
2025-01-13 18:56:31 -08:00
Patrick Devine
ab39872cb4
add new create api doc ( #8388 )
2025-01-13 17:30:24 -08:00
Parth Sareen
84a2314463
examples: remove codified examples ( #8267 )
2025-01-13 11:26:22 -08:00
Jeffrey Morgan
17fcdea698
readme: move discord link
2025-01-12 22:45:47 -08:00
Patrick Devine
32bd37adf8
make the modelfile path relative for ollama create
( #8380 )
v0.5.5
2025-01-10 16:14:08 -08:00
Michael Yang
9446c2c902
Merge pull request #8196 from ollama/mxyng/gods-v2
...
chore: upgrade to gods v2
2025-01-10 13:50:11 -08:00
Jeffrey Morgan
9aa141d023
readme: remove discord badge image for now
2025-01-09 22:02:18 -08:00
Patrick Devine
8bccae4f92
show a more descriptive error in the client if it is newer than the server ( #8351 )
2025-01-09 10:12:30 -08:00
isamu arimoto
6ae2adc1af
openai: accept additional headers to fix CORS errors ( #8343 )
v0.5.5-rc0
2025-01-08 11:28:11 -08:00
Jeffrey Morgan
1deafd8254
llama: update vendored code to commit 46e3556 ( #8308 )
2025-01-08 11:22:01 -08:00
Michael
57f038ec7b
readme: add phi4 model ( #8350 )
2025-01-08 11:21:39 -08:00
frob
cdf3a181dc
Add CUSTOM_CPU_FLAGS to Dockerfile. ( #8284 )
...
* Add CUSTOM_CPU_FLAGS.
* fix golangci-lint error.
---------
Co-authored-by: Richard Lyons <rick@frob.com.au>
2025-01-06 09:17:19 -08:00