mirror of
https://github.com/ollama/ollama.git
synced 2025-11-10 15:37:27 +01:00
* feat: Bump llama.cpp to df1b612 Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * fix(mtmd): Correctly encode text chunks during mtmd tokenization There can be text chunks that appear interspersed with the image embeddings that contain template delimiter tokens for some models. These need to be correctly translated to text tokens. Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * tests: Use MtmdChunk in image_test Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * style: Fix unnecessary conversion linting Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * fix(ggml): Revert changes to ggml_hip.cpp These changes were done largely by our code assistant and are likely wrong Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * fix: Revert changes in mem_nvml.cpp Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * feat: Update sync point to 1deee0 This brings in several more optimization commits and model support for EmbeddingGemma Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * feat: Update patches for 1deee0 Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * feat: sync for bump to 1deee0 Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * fix: Bad patch updates with errant `+` Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * feat: Bump llama.cpp/ggml to 7049736 Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> * fix: format-patches after latest bump Branch: LlamaCPPBump-GraniteDocling Signed-off-by: Gabe Goodhart <ghart@us.ibm.com> --------- Signed-off-by: Gabe Goodhart <ghart@us.ibm.com>
140 lines
6.1 KiB
Diff
140 lines
6.1 KiB
Diff
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
|
From: Jesse Gross <jesse@ollama.com>
|
|
Date: Fri, 18 Apr 2025 15:58:19 -0700
|
|
Subject: [PATCH] graph memory reporting on failure
|
|
|
|
---
|
|
ggml/include/ggml-alloc.h | 1 +
|
|
ggml/include/ggml-backend.h | 1 +
|
|
ggml/src/ggml-alloc.c | 34 +++++++++++++++++++++++++++++++---
|
|
ggml/src/ggml-backend.cpp | 7 +++++++
|
|
4 files changed, 40 insertions(+), 3 deletions(-)
|
|
|
|
diff --git a/ggml/include/ggml-alloc.h b/ggml/include/ggml-alloc.h
|
|
index 2cb150fd..7ab3f019 100644
|
|
--- a/ggml/include/ggml-alloc.h
|
|
+++ b/ggml/include/ggml-alloc.h
|
|
@@ -65,6 +65,7 @@ GGML_API bool ggml_gallocr_reserve_n(
|
|
GGML_API bool ggml_gallocr_alloc_graph(ggml_gallocr_t galloc, struct ggml_cgraph * graph);
|
|
|
|
GGML_API size_t ggml_gallocr_get_buffer_size(ggml_gallocr_t galloc, int buffer_id);
|
|
+GGML_API size_t ggml_gallocr_get_attempted_buffer_size(ggml_gallocr_t galloc, int buffer_id);
|
|
|
|
// Utils
|
|
// Create a buffer and allocate all the tensors in a ggml_context
|
|
diff --git a/ggml/include/ggml-backend.h b/ggml/include/ggml-backend.h
|
|
index f1b74078..c54ff98b 100644
|
|
--- a/ggml/include/ggml-backend.h
|
|
+++ b/ggml/include/ggml-backend.h
|
|
@@ -318,6 +318,7 @@ extern "C" {
|
|
|
|
GGML_API ggml_backend_buffer_type_t ggml_backend_sched_get_buffer_type(ggml_backend_sched_t sched, ggml_backend_t backend);
|
|
GGML_API size_t ggml_backend_sched_get_buffer_size(ggml_backend_sched_t sched, ggml_backend_t backend);
|
|
+ GGML_API size_t ggml_backend_sched_get_attempted_buffer_size(ggml_backend_sched_t sched, ggml_backend_t backend);
|
|
|
|
GGML_API void ggml_backend_sched_set_tensor_backend(ggml_backend_sched_t sched, struct ggml_tensor * node, ggml_backend_t backend);
|
|
GGML_API ggml_backend_t ggml_backend_sched_get_tensor_backend(ggml_backend_sched_t sched, struct ggml_tensor * node);
|
|
diff --git a/ggml/src/ggml-alloc.c b/ggml/src/ggml-alloc.c
|
|
index 929bc448..eee9d3b1 100644
|
|
--- a/ggml/src/ggml-alloc.c
|
|
+++ b/ggml/src/ggml-alloc.c
|
|
@@ -486,6 +486,7 @@ struct node_alloc {
|
|
struct ggml_gallocr {
|
|
ggml_backend_buffer_type_t * bufts; // [n_buffers]
|
|
struct vbuffer ** buffers; // [n_buffers]
|
|
+ size_t *buffer_sizes; // [n_buffers]
|
|
struct ggml_dyn_tallocr ** buf_tallocs; // [n_buffers]
|
|
int n_buffers;
|
|
|
|
@@ -509,6 +510,9 @@ ggml_gallocr_t ggml_gallocr_new_n(ggml_backend_buffer_type_t * bufts, int n_bufs
|
|
galloc->buffers = calloc(n_bufs, sizeof(struct vbuffer *));
|
|
GGML_ASSERT(galloc->buffers != NULL);
|
|
|
|
+ galloc->buffer_sizes = calloc(n_bufs, sizeof(size_t));
|
|
+ GGML_ASSERT(galloc->buffer_sizes != NULL);
|
|
+
|
|
galloc->buf_tallocs = calloc(n_bufs, sizeof(struct ggml_dyn_tallocr *));
|
|
GGML_ASSERT(galloc->buf_tallocs != NULL);
|
|
|
|
@@ -576,6 +580,7 @@ void ggml_gallocr_free(ggml_gallocr_t galloc) {
|
|
ggml_hash_set_free(&galloc->hash_set);
|
|
free(galloc->hash_values);
|
|
free(galloc->bufts);
|
|
+ free(galloc->buffer_sizes);
|
|
free(galloc->buffers);
|
|
free(galloc->buf_tallocs);
|
|
free(galloc->node_allocs);
|
|
@@ -869,6 +874,8 @@ bool ggml_gallocr_reserve_n(ggml_gallocr_t galloc, struct ggml_cgraph * graph, c
|
|
}
|
|
}
|
|
|
|
+ bool success = true;
|
|
+
|
|
// reallocate buffers if needed
|
|
for (int i = 0; i < galloc->n_buffers; i++) {
|
|
// if the buffer type is used multiple times, we reuse the same buffer
|
|
@@ -898,14 +905,19 @@ bool ggml_gallocr_reserve_n(ggml_gallocr_t galloc, struct ggml_cgraph * graph, c
|
|
|
|
ggml_vbuffer_free(galloc->buffers[i]);
|
|
galloc->buffers[i] = ggml_vbuffer_alloc(galloc->bufts[i], galloc->buf_tallocs[i], GGML_BACKEND_BUFFER_USAGE_COMPUTE);
|
|
- if (galloc->buffers[i] == NULL) {
|
|
+ if (galloc->buffers[i]) {
|
|
+ galloc->buffer_sizes[i] = ggml_vbuffer_size(galloc->buffers[i]);
|
|
+ } else {
|
|
GGML_LOG_ERROR("%s: failed to allocate %s buffer of size %zu\n", __func__, ggml_backend_buft_name(galloc->bufts[i]), new_size);
|
|
- return false;
|
|
+ galloc->buffer_sizes[i] = new_size;
|
|
+ success = false;
|
|
}
|
|
+ } else {
|
|
+ galloc->buffer_sizes[i] = ggml_vbuffer_size(galloc->buffers[i]);
|
|
}
|
|
}
|
|
|
|
- return true;
|
|
+ return success;
|
|
}
|
|
|
|
bool ggml_gallocr_reserve(ggml_gallocr_t galloc, struct ggml_cgraph *graph) {
|
|
@@ -1060,6 +1072,22 @@ size_t ggml_gallocr_get_buffer_size(ggml_gallocr_t galloc, int buffer_id) {
|
|
return ggml_vbuffer_size(galloc->buffers[buffer_id]);
|
|
}
|
|
|
|
+size_t ggml_gallocr_get_attempted_buffer_size(ggml_gallocr_t galloc, int buffer_id) {
|
|
+ GGML_ASSERT(buffer_id >= 0 && buffer_id < galloc->n_buffers);
|
|
+
|
|
+ for (int i = 0; i < buffer_id; i++) {
|
|
+ if (galloc->buf_tallocs[i] == galloc->buf_tallocs[buffer_id]) {
|
|
+ // This buffer is the same as a previous one due to the same buffer type being used multiple times
|
|
+ // (See above.) However, we need a different check because multiple buffers might be NULL in our
|
|
+ // case and we still want to know the attempted size.
|
|
+
|
|
+ return 0;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ return galloc->buffer_sizes[buffer_id];
|
|
+}
|
|
+
|
|
// utils
|
|
|
|
static void free_buffers(ggml_backend_buffer_t ** buffers, const size_t * n_buffers) {
|
|
diff --git a/ggml/src/ggml-backend.cpp b/ggml/src/ggml-backend.cpp
|
|
index 8ba86f82..cb2b9956 100644
|
|
--- a/ggml/src/ggml-backend.cpp
|
|
+++ b/ggml/src/ggml-backend.cpp
|
|
@@ -1809,6 +1809,13 @@ size_t ggml_backend_sched_get_buffer_size(ggml_backend_sched_t sched, ggml_backe
|
|
return ggml_gallocr_get_buffer_size(sched->galloc, backend_index);
|
|
}
|
|
|
|
+size_t ggml_backend_sched_get_attempted_buffer_size(ggml_backend_sched_t sched, ggml_backend_t backend) {
|
|
+ int backend_index = ggml_backend_sched_backend_id(sched, backend);
|
|
+ GGML_ASSERT(backend_index >= 0 && backend_index < sched->n_backends);
|
|
+
|
|
+ return ggml_gallocr_get_attempted_buffer_size(sched->galloc, backend_index);
|
|
+}
|
|
+
|
|
void ggml_backend_sched_set_tensor_backend(ggml_backend_sched_t sched, struct ggml_tensor * node, ggml_backend_t backend) {
|
|
GGML_ASSERT(sched);
|
|
int backend_index = ggml_backend_sched_backend_id(sched, backend);
|