7fab7918cc
Update import.md
2024-02-22 02:06:24 -05:00
f0425d3de9
Update faq.md
2024-02-20 20:44:45 -05:00
8125ce4cb6
Update import.md
...
Add instructions to get public key on windows
2024-02-19 22:48:24 -05:00
df56f1ee5e
Update faq.md
2024-02-19 22:16:42 -05:00
41aca5c2d0
Update faq.md
2024-02-19 21:11:01 -05:00
753724d867
Update api.md to include examples for reproducible outputs
2024-02-19 20:36:16 -05:00
9a7a4b9533
add faqs for memory pre-loading and the keep_alive setting ( #2601 )
2024-02-19 14:45:25 -08:00
b338c0635f
Document setting server vars for windows
2024-02-19 13:30:46 -08:00
9774663013
Update faq.md with the location of models on Windows ( #2545 )
2024-02-16 11:04:19 -08:00
1ba734de67
typo
2024-02-15 14:56:55 -08:00
29e90cc13b
Implement new Go based Desktop app
...
This focuses on Windows first, but coudl be used for Mac
and possibly linux in the future.
2024-02-15 05:56:45 +00:00
48a273f80b
Fix issues with templating prompt in chat mode ( #2460 )
2024-02-12 15:06:57 -08:00
1c8435ffa9
Update domain name references in docs and install script ( #2435 )
2024-02-09 15:19:30 -08:00
42b797ed9c
Update openai.md
2024-02-08 15:03:23 -05:00
336aa43f3c
Update openai.md
2024-02-08 12:48:28 -05:00
ab0d37fde4
Update openai.md
2024-02-07 17:25:33 -05:00
14e71350c8
Update openai.md
2024-02-07 17:25:24 -05:00
453f572f83
Initial OpenAI /v1/chat/completions
API compatibility ( #2376 )
2024-02-07 17:24:29 -05:00
128fce5495
docs: keep_alive ( #2258 )
2024-02-06 11:00:05 -05:00
b9f91a0b36
Update import instructions to use convert and quantize tooling from llama.cpp submodule ( #2247 )
2024-02-05 00:50:44 -05:00
f0e9496c85
Update api.md
2024-02-02 12:17:24 -08:00
e7dbb00331
Add container hints for troubleshooting
...
Some users are new to containers and unsure where the server logs go
2024-01-29 08:53:41 -08:00
e02ecfb6c8
Merge pull request #2116 from dhiltgen/cc_50_80
...
Add support for CUDA 5.0 cards
2024-01-27 10:28:38 -08:00
5be9bdd444
Update modelfile.md
2024-01-25 16:29:48 -08:00
b706794905
Update modelfile.md to include MESSAGE
2024-01-25 16:29:32 -08:00
93a756266c
faq: update to use launchctl setenv
2024-01-22 13:10:13 -08:00
df54c723ae
Make CPU builds parallel and customizable AMD GPUs
...
The linux build now support parallel CPU builds to speed things up.
This also exposes AMD GPU targets as an optional setting for advaced
users who want to alter our default set.
2024-01-21 15:12:21 -08:00
a447a083f2
Add compute capability 5.0, 7.5, and 8.0
2024-01-20 14:24:05 -08:00
abec7f06e5
Merge pull request #2056 from dhiltgen/slog
...
Mechanical switch from log to slog
2024-01-18 14:27:24 -08:00
ecbfc0182f
Go bump to v1.21 to pick up slog
2024-01-18 14:12:57 -08:00
fedd705aea
Mechanical switch from log to slog
...
A few obvious levels were adjusted, but generally everything mapped to "info" level.
2024-01-18 14:12:57 -08:00
9cd20b0ec8
Refine the linux cuda/rocm developer docs
2024-01-18 09:44:44 -08:00
40a0a90a88
Add group delete to uninstall instructions ( #1924 )
...
After executing the `userdel ollama` command, I saw this message:
```sh
$ sudo userdel ollama
userdel: group ollama not removed because it has other members.
```
Which reminded me that I had to remove the dangling group too. For completeness, the uninstall instructions should do this too.
Thanks!
2024-01-12 00:07:00 -05:00
d88c527be3
Build multiple CPU variants and pick the best
...
This reduces the built-in linux version to not use any vector extensions
which enables the resulting builds to run under Rosetta on MacOS in
Docker. Then at runtime it checks for the actual CPU vector
extensions and loads the best CPU library available
2024-01-11 08:42:47 -08:00
e868c8a5c7
Update api.md ( #1878 )
...
Fixed assistant in the example response.
2024-01-09 16:21:17 -05:00
3f3eb19a3b
document response in modelfile template variables ( #1428 )
2024-01-08 14:38:51 -05:00
2d9dd14f27
Merge pull request #1697 from dhiltgen/win_docs
...
Add windows native build instructions
2024-01-05 19:34:20 -08:00
df086d3c8c
fix docker doc to point to hub
...
Signed-off-by: Matt Williams <m@technovangelist.com >
2024-01-04 18:42:23 -08:00
b846eb64d0
Fix template
api doc description ( #1661 )
2024-01-03 11:00:59 -05:00
3c5dd9ed1d
Update README.md ( #1766 )
2024-01-03 10:44:22 -05:00
b17ccd0542
Update import.md
2024-01-02 22:28:18 -05:00
2a2fa3c329
api.md
cleanup & formatting
2023-12-27 14:32:35 -05:00
e201efa14b
Add windows native build instructions
2023-12-25 08:31:34 -08:00
10da41d677
Add Cache flag to api ( #1642 )
2023-12-22 17:16:20 -05:00
511069a2a5
update where are models stored q
...
Signed-off-by: Matt Williams <m@technovangelist.com >
2023-12-22 09:48:44 -08:00
291700c92d
Clean up documentation ( #1506 )
...
* Clean up documentation
Will probably need to update with PRs for new release.
Signed-off-by: Matt Williams <m@technovangelist.com >
* Correcting to fit in 0.1.15 changes
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* addressing comments
Signed-off-by: Matt Williams <m@technovangelist.com >
* more api cleanup
Signed-off-by: Matt Williams <m@technovangelist.com >
* its llava not llama
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update docs/troubleshooting.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Updated hosting to server and documented all env vars
Signed-off-by: Matt Williams <m@technovangelist.com >
* remove last of the cli descriptions
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* update further per conversation with jeff earlier today
Signed-off-by: Matt Williams <m@technovangelist.com >
* cleanup the doc readme
Signed-off-by: Matt Williams <m@technovangelist.com >
* move upgrade to faq
Signed-off-by: Matt Williams <m@technovangelist.com >
* first change
Signed-off-by: Matt Williams <m@technovangelist.com >
* updated
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update docs/faq.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* examples in parent
Signed-off-by: Matt Williams <m@technovangelist.com >
* add exapmle for create model.
Signed-off-by: Matt Williams <m@technovangelist.com >
* update faq
Signed-off-by: Matt Williams <m@technovangelist.com >
* update create model api
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update docs/api.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/faq.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/troubleshooting.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* update the readme in docs
Signed-off-by: Matt Williams <m@technovangelist.com >
* update a few more things
Signed-off-by: Matt Williams <m@technovangelist.com >
* Update docs/troubleshooting.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/faq.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update README.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/modelfile.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
* Update docs/troubleshooting.md
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
---------
Signed-off-by: Matt Williams <m@technovangelist.com >
Co-authored-by: Jeffrey Morgan <jmorganca@gmail.com >
2023-12-22 09:10:01 -08:00
e5202eb687
Quiet down llama.cpp logging by default
...
By default builds will now produce non-debug and non-verbose binaries.
To enable verbose logs in llama.cpp and debug symbols in the
native code, set `CGO_CFLAGS=-g`
2023-12-22 08:47:18 -08:00
96fb441abd
Merge pull request #1146 from dhiltgen/ext_server_cgo
...
Add cgo implementation for llama.cpp
2023-12-22 08:16:31 -08:00
495c06e4a6
Fix doc glitch
2023-12-21 18:21:31 -08:00
a607d922f0
add FAQ for slow networking in WSL2 ( #1646 )
2023-12-20 16:27:24 -08:00