diff --git a/.golangci.yaml b/.golangci.yaml index 85558f035..985fad639 100644 --- a/.golangci.yaml +++ b/.golangci.yaml @@ -24,4 +24,4 @@ linters-settings: - (*os.File).Seek - (*bufio.Writer).WriteString - (*github.com/spf13/pflag.FlagSet).Set - - (*github.com/jmorganca/ollama/llm.readSeekOffset).Seek + - (*github.com/ollama/ollama/llm.readSeekOffset).Seek diff --git a/Dockerfile b/Dockerfile index 59cca725a..919d48de1 100644 --- a/Dockerfile +++ b/Dockerfile @@ -15,8 +15,8 @@ ARG CMAKE_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh @@ -25,8 +25,8 @@ ARG CMAKE_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/gcc-toolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh @@ -36,18 +36,18 @@ COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH ENV LIBRARY_PATH /opt/amdgpu/lib64 -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS ARG AMDGPU_TARGETS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh RUN mkdir /tmp/scratch && \ - for dep in $(cat /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/x86_64/rocm*/lib/deps.txt) ; do \ + for dep in $(cat /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/x86_64/rocm*/lib/deps.txt) ; do \ cp ${dep} /tmp/scratch/ || exit 1 ; \ done && \ (cd /opt/rocm/lib && tar cf - rocblas/library) | (cd /tmp/scratch/ && tar xf - ) && \ - mkdir -p /go/src/github.com/jmorganca/ollama/dist/deps/ && \ - (cd /tmp/scratch/ && tar czvf /go/src/github.com/jmorganca/ollama/dist/deps/ollama-linux-amd64-rocm.tgz . ) + mkdir -p /go/src/github.com/ollama/ollama/dist/deps/ && \ + (cd /tmp/scratch/ && tar czvf /go/src/github.com/ollama/ollama/dist/deps/ollama-linux-amd64-rocm.tgz . ) FROM --platform=linux/amd64 centos:7 AS cpu-builder-amd64 @@ -56,10 +56,10 @@ ARG GOLANG_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} GOLANG_VERSION=${GOLANG_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ ARG OLLAMA_CUSTOM_CPU_DEFS ARG CGO_CFLAGS -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +WORKDIR /go/src/github.com/ollama/ollama/llm/generate FROM --platform=linux/amd64 cpu-builder-amd64 AS cpu-build-amd64 RUN OLLAMA_CPU_TARGET="cpu" sh gen_linux.sh @@ -74,8 +74,8 @@ ARG GOLANG_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} GOLANG_VERSION=${GOLANG_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate # Note, we only build the "base" CPU variant on arm since avx/avx2 are x86 features ARG OLLAMA_CUSTOM_CPU_DEFS ARG CGO_CFLAGS @@ -84,13 +84,13 @@ RUN OLLAMA_CPU_TARGET="cpu" sh gen_linux.sh # Intermediate stage used for ./scripts/build_linux.sh FROM --platform=linux/amd64 cpu-build-amd64 AS build-amd64 ENV CGO_ENABLED 1 -WORKDIR /go/src/github.com/jmorganca/ollama +WORKDIR /go/src/github.com/ollama/ollama COPY . . -COPY --from=cpu_avx-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=cpu_avx2-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=cuda-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=rocm-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=rocm-build-amd64 /go/src/github.com/jmorganca/ollama/dist/deps/ ./dist/deps/ +COPY --from=cpu_avx-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=cpu_avx2-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=cuda-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=rocm-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=rocm-build-amd64 /go/src/github.com/ollama/ollama/dist/deps/ ./dist/deps/ ARG GOFLAGS ARG CGO_CFLAGS RUN go build -trimpath . @@ -99,10 +99,10 @@ RUN go build -trimpath . FROM --platform=linux/arm64 cpu-build-arm64 AS build-arm64 ENV CGO_ENABLED 1 ARG GOLANG_VERSION -WORKDIR /go/src/github.com/jmorganca/ollama +WORKDIR /go/src/github.com/ollama/ollama COPY . . -COPY --from=cuda-build-arm64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -RUN mkdir -p /go/src/github.com/jmorganca/ollama/dist/deps/ +COPY --from=cuda-build-arm64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +RUN mkdir -p /go/src/github.com/ollama/ollama/dist/deps/ ARG GOFLAGS ARG CGO_CFLAGS RUN go build -trimpath . @@ -110,15 +110,15 @@ RUN go build -trimpath . # Runtime stages FROM --platform=linux/amd64 ubuntu:22.04 as runtime-amd64 RUN apt-get update && apt-get install -y ca-certificates -COPY --from=build-amd64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-amd64 /go/src/github.com/ollama/ollama/ollama /bin/ollama FROM --platform=linux/arm64 ubuntu:22.04 as runtime-arm64 RUN apt-get update && apt-get install -y ca-certificates -COPY --from=build-arm64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-arm64 /go/src/github.com/ollama/ollama/ollama /bin/ollama # Radeon images are much larger so we keep it distinct from the CPU/CUDA image FROM --platform=linux/amd64 rocm/dev-centos-7:${ROCM_VERSION}-complete as runtime-rocm RUN update-pciids -COPY --from=build-amd64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-amd64 /go/src/github.com/ollama/ollama/ollama /bin/ollama EXPOSE 11434 ENV OLLAMA_HOST 0.0.0.0 diff --git a/README.md b/README.md index 48da0e96c..f46e479fe 100644 --- a/README.md +++ b/README.md @@ -1,5 +1,5 @@
- ollama + ollama
# Ollama @@ -22,7 +22,7 @@ Get up and running with large language models locally. curl -fsSL https://ollama.com/install.sh | sh ``` -[Manual install instructions](https://github.com/jmorganca/ollama/blob/main/docs/linux.md) +[Manual install instructions](https://github.com/ollama/ollama/blob/main/docs/linux.md) ### Docker @@ -213,7 +213,7 @@ Then build the binary: go build . ``` -More detailed instructions can be found in the [developer guide](https://github.com/jmorganca/ollama/blob/main/docs/development.md) +More detailed instructions can be found in the [developer guide](https://github.com/ollama/ollama/blob/main/docs/development.md) ### Running local builds diff --git a/api/client.go b/api/client.go index 36019206a..a37a4cbbd 100644 --- a/api/client.go +++ b/api/client.go @@ -15,8 +15,8 @@ import ( "runtime" "strings" - "github.com/jmorganca/ollama/format" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/format" + "github.com/ollama/ollama/version" ) type Client struct { diff --git a/app/lifecycle/lifecycle.go b/app/lifecycle/lifecycle.go index 14a85b11c..ab624e817 100644 --- a/app/lifecycle/lifecycle.go +++ b/app/lifecycle/lifecycle.go @@ -9,8 +9,8 @@ import ( "os/signal" "syscall" - "github.com/jmorganca/ollama/app/store" - "github.com/jmorganca/ollama/app/tray" + "github.com/ollama/ollama/app/store" + "github.com/ollama/ollama/app/tray" ) func Run() { diff --git a/app/lifecycle/server.go b/app/lifecycle/server.go index 1cb689a20..e3ca22f97 100644 --- a/app/lifecycle/server.go +++ b/app/lifecycle/server.go @@ -11,7 +11,7 @@ import ( "path/filepath" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func getCLIFullPath(command string) string { diff --git a/app/lifecycle/updater.go b/app/lifecycle/updater.go index f26e32af1..243bbf223 100644 --- a/app/lifecycle/updater.go +++ b/app/lifecycle/updater.go @@ -18,8 +18,8 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/auth" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/auth" + "github.com/ollama/ollama/version" ) var ( diff --git a/app/main.go b/app/main.go index 57d8b1c14..db8297954 100644 --- a/app/main.go +++ b/app/main.go @@ -4,7 +4,7 @@ package main // go build -ldflags="-H windowsgui" . import ( - "github.com/jmorganca/ollama/app/lifecycle" + "github.com/ollama/ollama/app/lifecycle" ) func main() { diff --git a/app/tray/tray.go b/app/tray/tray.go index 47b204d62..20d1c0b54 100644 --- a/app/tray/tray.go +++ b/app/tray/tray.go @@ -4,8 +4,8 @@ import ( "fmt" "runtime" - "github.com/jmorganca/ollama/app/assets" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/assets" + "github.com/ollama/ollama/app/tray/commontray" ) func NewTray() (commontray.OllamaTray, error) { diff --git a/app/tray/tray_nonwindows.go b/app/tray/tray_nonwindows.go index 6c30c3c25..ae5572b23 100644 --- a/app/tray/tray_nonwindows.go +++ b/app/tray/tray_nonwindows.go @@ -5,7 +5,7 @@ package tray import ( "fmt" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/commontray" ) func InitPlatformTray(icon, updateIcon []byte) (commontray.OllamaTray, error) { diff --git a/app/tray/tray_windows.go b/app/tray/tray_windows.go index 8ac4e4788..086fc7943 100644 --- a/app/tray/tray_windows.go +++ b/app/tray/tray_windows.go @@ -1,8 +1,8 @@ package tray import ( - "github.com/jmorganca/ollama/app/tray/commontray" - "github.com/jmorganca/ollama/app/tray/wintray" + "github.com/ollama/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/wintray" ) func InitPlatformTray(icon, updateIcon []byte) (commontray.OllamaTray, error) { diff --git a/app/tray/wintray/tray.go b/app/tray/wintray/tray.go index 365cfb823..69d4487de 100644 --- a/app/tray/wintray/tray.go +++ b/app/tray/wintray/tray.go @@ -13,7 +13,7 @@ import ( "sync" "unsafe" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/commontray" "golang.org/x/sys/windows" ) diff --git a/cmd/cmd.go b/cmd/cmd.go index 3a0419e40..c5430f6f5 100644 --- a/cmd/cmd.go +++ b/cmd/cmd.go @@ -30,12 +30,12 @@ import ( "golang.org/x/exp/slices" "golang.org/x/term" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/progress" - "github.com/jmorganca/ollama/server" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/progress" + "github.com/ollama/ollama/server" + "github.com/ollama/ollama/version" ) func CreateHandler(cmd *cobra.Command, args []string) error { diff --git a/cmd/interactive.go b/cmd/interactive.go index 82e3642a3..2dd9ee194 100644 --- a/cmd/interactive.go +++ b/cmd/interactive.go @@ -14,9 +14,9 @@ import ( "github.com/spf13/cobra" "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/progress" - "github.com/jmorganca/ollama/readline" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/progress" + "github.com/ollama/ollama/readline" ) type MultilineState int diff --git a/cmd/interactive_test.go b/cmd/interactive_test.go index 19e43287e..8eedf7292 100644 --- a/cmd/interactive_test.go +++ b/cmd/interactive_test.go @@ -7,7 +7,7 @@ import ( "github.com/stretchr/testify/assert" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestExtractFilenames(t *testing.T) { diff --git a/cmd/start_darwin.go b/cmd/start_darwin.go index 7e3000f09..82b09ad62 100644 --- a/cmd/start_darwin.go +++ b/cmd/start_darwin.go @@ -7,7 +7,7 @@ import ( "os/exec" "strings" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/cmd/start_default.go b/cmd/start_default.go index 664c2d1f2..c9d6137b9 100644 --- a/cmd/start_default.go +++ b/cmd/start_default.go @@ -6,7 +6,7 @@ import ( "context" "fmt" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/cmd/start_windows.go b/cmd/start_windows.go index b9a423cfb..6024a2352 100644 --- a/cmd/start_windows.go +++ b/cmd/start_windows.go @@ -10,7 +10,7 @@ import ( "strings" "syscall" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/convert/convert.go b/convert/convert.go index ba23080c5..4768d6f22 100644 --- a/convert/convert.go +++ b/convert/convert.go @@ -16,8 +16,8 @@ import ( "github.com/mitchellh/mapstructure" "google.golang.org/protobuf/proto" - "github.com/jmorganca/ollama/convert/sentencepiece" - "github.com/jmorganca/ollama/llm" + "github.com/ollama/ollama/convert/sentencepiece" + "github.com/ollama/ollama/llm" ) type Params struct { diff --git a/docs/faq.md b/docs/faq.md index 63976ee45..6bd1b3404 100644 --- a/docs/faq.md +++ b/docs/faq.md @@ -154,7 +154,7 @@ No. Ollama runs locally, and conversation data does not leave your machine. ## How can I use Ollama in Visual Studio Code? -There is already a large collection of plugins available for VSCode as well as other editors that leverage Ollama. See the list of [extensions & plugins](https://github.com/jmorganca/ollama#extensions--plugins) at the bottom of the main repository readme. +There is already a large collection of plugins available for VSCode as well as other editors that leverage Ollama. See the list of [extensions & plugins](https://github.com/ollama/ollama#extensions--plugins) at the bottom of the main repository readme. ## How do I use Ollama behind a proxy? diff --git a/docs/modelfile.md b/docs/modelfile.md index 7edd85eb4..94acbdf6c 100644 --- a/docs/modelfile.md +++ b/docs/modelfile.md @@ -113,7 +113,7 @@ FROM llama2 ``` A list of available base models: - + #### Build from a `bin` file diff --git a/docs/openai.md b/docs/openai.md index 5808ae756..b4dc1f21f 100644 --- a/docs/openai.md +++ b/docs/openai.md @@ -1,6 +1,6 @@ # OpenAI compatibility -> **Note:** OpenAI compatibility is experimental and is subject to major adjustments including breaking changes. For fully-featured access to the Ollama API, see the Ollama [Python library](https://github.com/ollama/ollama-python), [JavaScript library](https://github.com/ollama/ollama-js) and [REST API](https://github.com/jmorganca/ollama/blob/main/docs/api.md). +> **Note:** OpenAI compatibility is experimental and is subject to major adjustments including breaking changes. For fully-featured access to the Ollama API, see the Ollama [Python library](https://github.com/ollama/ollama-python), [JavaScript library](https://github.com/ollama/ollama-js) and [REST API](https://github.com/ollama/ollama/blob/main/docs/api.md). Ollama provides experimental compatibility with parts of the [OpenAI API](https://platform.openai.com/docs/api-reference) to help connect existing applications to Ollama. diff --git a/examples/langchain-python-rag-privategpt/README.md b/examples/langchain-python-rag-privategpt/README.md index f739df2b5..0133fd88b 100644 --- a/examples/langchain-python-rag-privategpt/README.md +++ b/examples/langchain-python-rag-privategpt/README.md @@ -1,6 +1,6 @@ # PrivateGPT with Llama 2 uncensored -https://github.com/jmorganca/ollama/assets/3325447/20cf8ec6-ff25-42c6-bdd8-9be594e3ce1b +https://github.com/ollama/ollama/assets/3325447/20cf8ec6-ff25-42c6-bdd8-9be594e3ce1b > Note: this example is a slightly modified version of PrivateGPT using models such as Llama 2 Uncensored. All credit for PrivateGPT goes to Iván Martínez who is the creator of it, and you can find his GitHub repo [here](https://github.com/imartinez/privateGPT). diff --git a/examples/modelfile-mario/readme.md b/examples/modelfile-mario/readme.md index 9df08a675..0d72dddcf 100644 --- a/examples/modelfile-mario/readme.md +++ b/examples/modelfile-mario/readme.md @@ -28,7 +28,7 @@ You are Mario from Super Mario Bros, acting as an assistant. What if you want to change its behaviour? - Try changing the prompt -- Try changing the parameters [Docs](https://github.com/jmorganca/ollama/blob/main/docs/modelfile.md) +- Try changing the parameters [Docs](https://github.com/ollama/ollama/blob/main/docs/modelfile.md) - Try changing the model (e.g. An uncensored model by `FROM wizard-vicuna` this is the wizard-vicuna uncensored model ) Once the changes are made, diff --git a/examples/python-json-datagenerator/readme.md b/examples/python-json-datagenerator/readme.md index ec5701bea..369fb2a5b 100644 --- a/examples/python-json-datagenerator/readme.md +++ b/examples/python-json-datagenerator/readme.md @@ -1,6 +1,6 @@ # JSON Output Example -![llmjson 2023-11-10 15_31_31](https://github.com/jmorganca/ollama/assets/633681/e599d986-9b4a-4118-81a4-4cfe7e22da25) +![llmjson 2023-11-10 15_31_31](https://github.com/ollama/ollama/assets/633681/e599d986-9b4a-4118-81a4-4cfe7e22da25) There are two python scripts in this example. `randomaddresses.py` generates random addresses from different countries. `predefinedschema.py` sets a template for the model to fill in. diff --git a/examples/python-loganalysis/readme.md b/examples/python-loganalysis/readme.md index 60c57217d..4be0baaaf 100644 --- a/examples/python-loganalysis/readme.md +++ b/examples/python-loganalysis/readme.md @@ -1,6 +1,6 @@ # Log Analysis example -![loganalyzer 2023-11-10 08_53_29](https://github.com/jmorganca/ollama/assets/633681/ad30f1fc-321f-4953-8914-e30e24db9921) +![loganalyzer 2023-11-10 08_53_29](https://github.com/ollama/ollama/assets/633681/ad30f1fc-321f-4953-8914-e30e24db9921) This example shows one possible way to create a log file analyzer. It uses the model **mattw/loganalyzer** which is based on **codebooga**, a 34b parameter model. diff --git a/examples/typescript-functioncalling/readme.md b/examples/typescript-functioncalling/readme.md index 03ff46dcf..d29379a03 100644 --- a/examples/typescript-functioncalling/readme.md +++ b/examples/typescript-functioncalling/readme.md @@ -1,6 +1,6 @@ # Function calling -![function calling 2023-11-16 16_12_58](https://github.com/jmorganca/ollama/assets/633681/a0acc247-9746-45ab-b325-b65dfbbee4fb) +![function calling 2023-11-16 16_12_58](https://github.com/ollama/ollama/assets/633681/a0acc247-9746-45ab-b325-b65dfbbee4fb) One of the features added to some models is 'function calling'. It's a bit of a confusing name. It's understandable if you think that means the model can call functions, but that's not what it means. Function calling simply means that the output of the model is formatted in JSON, using a preconfigured schema, and uses the expected types. Then your code can use the output of the model and call functions with it. Using the JSON format in Ollama, you can use any model for function calling. diff --git a/go.mod b/go.mod index 74f75b47a..65647fb14 100644 --- a/go.mod +++ b/go.mod @@ -1,4 +1,4 @@ -module github.com/jmorganca/ollama +module github.com/ollama/ollama go 1.22 diff --git a/integration/basic_test.go b/integration/basic_test.go index ce933ffef..926ca52c9 100644 --- a/integration/basic_test.go +++ b/integration/basic_test.go @@ -8,7 +8,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestOrcaMiniBlueSky(t *testing.T) { diff --git a/integration/llm_image_test.go b/integration/llm_image_test.go index 0ac0e1e80..94082d6e0 100644 --- a/integration/llm_image_test.go +++ b/integration/llm_image_test.go @@ -9,7 +9,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" "github.com/stretchr/testify/require" ) diff --git a/integration/llm_test.go b/integration/llm_test.go index 107b5573a..5b0604478 100644 --- a/integration/llm_test.go +++ b/integration/llm_test.go @@ -9,7 +9,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) // TODO - this would ideally be in the llm package, but that would require some refactoring of interfaces in the server diff --git a/integration/utils_test.go b/integration/utils_test.go index 47184af89..14ec39dce 100644 --- a/integration/utils_test.go +++ b/integration/utils_test.go @@ -21,8 +21,8 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/app/lifecycle" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/app/lifecycle" "github.com/stretchr/testify/assert" ) diff --git a/llm/dyn_ext_server.go b/llm/dyn_ext_server.go index a6a010d7c..6e43333ee 100644 --- a/llm/dyn_ext_server.go +++ b/llm/dyn_ext_server.go @@ -33,8 +33,8 @@ import ( "time" "unsafe" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" ) type dynExtServer struct { diff --git a/llm/gguf.go b/llm/gguf.go index 0c7e914f2..6cb90484f 100644 --- a/llm/gguf.go +++ b/llm/gguf.go @@ -15,7 +15,7 @@ import ( "github.com/pdevine/tensor/native" "github.com/x448/float16" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/format" ) type ContainerGGUF struct { diff --git a/llm/llama.go b/llm/llama.go index a5d2036a1..cce9f4842 100644 --- a/llm/llama.go +++ b/llm/llama.go @@ -5,7 +5,7 @@ import ( "fmt" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) const jsonGrammar = ` diff --git a/llm/llm.go b/llm/llm.go index a3f59d2ed..75ff955eb 100644 --- a/llm/llm.go +++ b/llm/llm.go @@ -8,8 +8,8 @@ import ( "runtime" "slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" ) type LLM interface { diff --git a/llm/payload_common.go b/llm/payload_common.go index d45b70c47..45e45a2dc 100644 --- a/llm/payload_common.go +++ b/llm/payload_common.go @@ -16,7 +16,7 @@ import ( "golang.org/x/exp/slices" "golang.org/x/sync/errgroup" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/gpu" ) // Libraries names may contain an optional variant separated by '_' diff --git a/llm/payload_test.go b/llm/payload_test.go index 44537b0a3..eb88d8125 100644 --- a/llm/payload_test.go +++ b/llm/payload_test.go @@ -3,7 +3,7 @@ package llm import ( "testing" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/gpu" "github.com/stretchr/testify/assert" ) diff --git a/main.go b/main.go index ecdcb71db..158f0063c 100644 --- a/main.go +++ b/main.go @@ -3,7 +3,7 @@ package main import ( "context" - "github.com/jmorganca/ollama/cmd" + "github.com/ollama/ollama/cmd" "github.com/spf13/cobra" ) diff --git a/openai/openai.go b/openai/openai.go index 4f495569e..96d8f218a 100644 --- a/openai/openai.go +++ b/openai/openai.go @@ -11,7 +11,7 @@ import ( "time" "github.com/gin-gonic/gin" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) type Error struct { diff --git a/progress/bar.go b/progress/bar.go index 9a0da8256..476ea8881 100644 --- a/progress/bar.go +++ b/progress/bar.go @@ -6,7 +6,7 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/format" "golang.org/x/term" ) diff --git a/scripts/build_darwin.sh b/scripts/build_darwin.sh index cb5611236..a2f76af2f 100755 --- a/scripts/build_darwin.sh +++ b/scripts/build_darwin.sh @@ -3,7 +3,7 @@ set -e export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" mkdir -p dist diff --git a/scripts/build_docker.sh b/scripts/build_docker.sh index 7bdad6412..958733f32 100755 --- a/scripts/build_docker.sh +++ b/scripts/build_docker.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" # We use 2 different image repositories to handle combining architecture images into multiarch manifest # (The ROCm image is x86 only and is not a multiarch manifest) @@ -74,4 +74,4 @@ if [ -z "${OLLAMA_SKIP_MANIFEST_CREATE}" ]; then echo " ${ARCH_IMAGE_REPO}:$VERSION-arm64" echo " ${ARCH_IMAGE_REPO}:$VERSION-rocm" fi -fi \ No newline at end of file +fi diff --git a/scripts/build_linux.sh b/scripts/build_linux.sh index e6db485af..27c4ff1f8 100755 --- a/scripts/build_linux.sh +++ b/scripts/build_linux.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" BUILD_ARCH=${BUILD_ARCH:-"amd64 arm64"} export AMDGPU_TARGETS=${AMDGPU_TARGETS:=""} @@ -21,10 +21,10 @@ for TARGETARCH in ${BUILD_ARCH}; do -t builder:$TARGETARCH \ . docker create --platform linux/$TARGETARCH --name builder-$TARGETARCH builder:$TARGETARCH - docker cp builder-$TARGETARCH:/go/src/github.com/jmorganca/ollama/ollama ./dist/ollama-linux-$TARGETARCH + docker cp builder-$TARGETARCH:/go/src/github.com/ollama/ollama/ollama ./dist/ollama-linux-$TARGETARCH if [ "$TARGETARCH" = "amd64" ]; then - docker cp builder-$TARGETARCH:/go/src/github.com/jmorganca/ollama/dist/deps/ ./dist/ + docker cp builder-$TARGETARCH:/go/src/github.com/ollama/ollama/dist/deps/ ./dist/ fi docker rm builder-$TARGETARCH diff --git a/scripts/build_windows.ps1 b/scripts/build_windows.ps1 index e6e1f4b0e..1a89045aa 100644 --- a/scripts/build_windows.ps1 +++ b/scripts/build_windows.ps1 @@ -74,7 +74,7 @@ function buildOllama() { } else { write-host "Skipping generate step with OLLAMA_SKIP_GENERATE set" } - & go build -trimpath -ldflags "-s -w -X=github.com/jmorganca/ollama/version.Version=$script:VERSION -X=github.com/jmorganca/ollama/server.mode=release" . + & go build -trimpath -ldflags "-s -w -X=github.com/ollama/ollama/version.Version=$script:VERSION -X=github.com/ollama/ollama/server.mode=release" . if ($LASTEXITCODE -ne 0) { exit($LASTEXITCODE)} if ("${env:KEY_CONTAINER}") { & "${script:SignTool}" sign /v /fd sha256 /t http://timestamp.digicert.com /f "${script:OLLAMA_CERT}" ` @@ -89,7 +89,7 @@ function buildApp() { write-host "Building Ollama App" cd "${script:SRC_DIR}\app" & windres -l 0 -o ollama.syso ollama.rc - & go build -trimpath -ldflags "-s -w -H windowsgui -X=github.com/jmorganca/ollama/version.Version=$script:VERSION -X=github.com/jmorganca/ollama/server.mode=release" . + & go build -trimpath -ldflags "-s -w -H windowsgui -X=github.com/ollama/ollama/version.Version=$script:VERSION -X=github.com/ollama/ollama/server.mode=release" . if ($LASTEXITCODE -ne 0) { exit($LASTEXITCODE)} if ("${env:KEY_CONTAINER}") { & "${script:SignTool}" sign /v /fd sha256 /t http://timestamp.digicert.com /f "${script:OLLAMA_CERT}" ` diff --git a/scripts/push_docker.sh b/scripts/push_docker.sh index da953dd48..c140c27c7 100755 --- a/scripts/push_docker.sh +++ b/scripts/push_docker.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-0.0.0} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" docker build \ --push \ diff --git a/server/auth.go b/server/auth.go index 5af85ff65..e92a5b657 100644 --- a/server/auth.go +++ b/server/auth.go @@ -15,8 +15,8 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/auth" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/auth" ) type registryChallenge struct { diff --git a/server/download.go b/server/download.go index f6d199b92..db9b9681a 100644 --- a/server/download.go +++ b/server/download.go @@ -21,8 +21,8 @@ import ( "golang.org/x/sync/errgroup" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" ) const maxRetries = 6 diff --git a/server/images.go b/server/images.go index 11c117450..e5837d917 100644 --- a/server/images.go +++ b/server/images.go @@ -24,11 +24,11 @@ import ( "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/convert" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/convert" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) type registryOptions struct { diff --git a/server/prompt.go b/server/prompt.go index 88da5b6bc..604e69717 100644 --- a/server/prompt.go +++ b/server/prompt.go @@ -7,7 +7,7 @@ import ( "text/template" "text/template/parse" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) // isResponseNode checks if the node contains .Response diff --git a/server/prompt_test.go b/server/prompt_test.go index 500ee5225..a7e18a70f 100644 --- a/server/prompt_test.go +++ b/server/prompt_test.go @@ -4,7 +4,7 @@ import ( "strings" "testing" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestPrompt(t *testing.T) { diff --git a/server/routes.go b/server/routes.go index a03f39e77..5582f3a15 100644 --- a/server/routes.go +++ b/server/routes.go @@ -27,12 +27,12 @@ import ( "github.com/gin-gonic/gin" "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/openai" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/openai" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) var mode string = gin.DebugMode diff --git a/server/routes_test.go b/server/routes_test.go index bbed02ede..9b2989701 100644 --- a/server/routes_test.go +++ b/server/routes_test.go @@ -15,10 +15,10 @@ import ( "github.com/stretchr/testify/assert" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) func Test_Routes(t *testing.T) { diff --git a/server/upload.go b/server/upload.go index 4da34052a..9b52238a4 100644 --- a/server/upload.go +++ b/server/upload.go @@ -16,8 +16,8 @@ import ( "sync/atomic" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" "golang.org/x/sync/errgroup" )