mirror of
https://github.com/ollama/ollama.git
synced 2025-03-18 13:51:45 +01:00
61 lines
1.7 KiB
Makefile
61 lines
1.7 KiB
Makefile
UPSTREAM=https://github.com/ggerganov/llama.cpp.git
|
|
WORKDIR=llama/vendor
|
|
FETCH_HEAD=d7cfe1ffe0f435d0048a6058d529daf76e072d9c
|
|
|
|
.PHONY: help
|
|
help:
|
|
@echo "Available targets:"
|
|
@echo " sync Sync with upstream repositories"
|
|
@echo " checkout Checkout upstream repository"
|
|
@echo " apply-patches Apply patches to local repository"
|
|
@echo " format-patches Format patches from local repository"
|
|
@echo " clean Clean local repository"
|
|
@echo
|
|
@echo "Example:"
|
|
@echo " make -f $(lastword $(MAKEFILE_LIST)) clean sync"
|
|
|
|
.PHONY: sync
|
|
sync: llama/build-info.cpp llama/llama.cpp ml/backend/ggml/ggml apply-patches
|
|
|
|
.PHONY: llama/build-info.cpp
|
|
llama/build-info.cpp: llama/build-info.cpp.in
|
|
sed -e 's|@FETCH_HEAD@|$(FETCH_HEAD)|' $< > $@
|
|
|
|
.PHONY: llama/llama.cpp
|
|
llama/llama.cpp: llama/vendor/ apply-patches
|
|
rsync -arvzc -f "merge $@/.rsync-filter" $< $@
|
|
|
|
.PHONY: ml/backend/ggml/ggml apply-patches
|
|
ml/backend/ggml/ggml: llama/vendor/ggml/ apply-patches
|
|
rsync -arvzc -f "merge $@/.rsync-filter" $< $@
|
|
|
|
PATCHES=$(wildcard llama/patches/*.patch)
|
|
|
|
.PHONY: apply-patches
|
|
.NOTPARALLEL:
|
|
apply-patches: $(addsuffix ed, $(PATCHES))
|
|
|
|
%.patched: %.patch
|
|
@if git -c user.name=nobody -c 'user.email=<>' -C $(WORKDIR) am -3 $(realpath $<); then touch $@; else git -C $(WORKDIR) am --abort; exit 1; fi
|
|
|
|
.PHONY: checkout
|
|
checkout: $(WORKDIR)
|
|
git -C $(WORKDIR) fetch
|
|
git -C $(WORKDIR) checkout -f $(FETCH_HEAD)
|
|
|
|
$(WORKDIR):
|
|
git clone $(UPSTREAM) $(WORKDIR)
|
|
|
|
.PHONE: format-patches
|
|
format-patches: llama/patches
|
|
git -C $(WORKDIR) format-patch \
|
|
--no-signature \
|
|
--no-numbered \
|
|
--zero-commit \
|
|
-o $(realpath $<) \
|
|
$(FETCH_HEAD)
|
|
|
|
.PHONE: clean
|
|
clean: checkout
|
|
$(RM) $(addsuffix ed, $(PATCHES))
|