mirror of https://github.com/ollama/ollama.git
				
				
				
			llama: fix kv loading on snowflake-arctic-embed models (#9536)
This commit is contained in:
		
							parent
							
								
									a7e63b82be
								
							
						
					
					
						commit
						4289c74359
					
				|  | @ -1443,7 +1443,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) { | |||
| 
 | ||||
|             const int precompiled_charsmap_keyidx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_PRECOMPILED_CHARSMAP).c_str()); | ||||
|             if (precompiled_charsmap_keyidx != -1) { | ||||
|                 size_t n_precompiled_charsmap = gguf_get_arr_n(ctx, precompiled_charsmap_keyidx); | ||||
|                 size_t n_precompiled_charsmap = gguf_get_arr_data_n(ctx, precompiled_charsmap_keyidx); | ||||
|                 const char * pc = (const char *) gguf_get_arr_data(ctx, precompiled_charsmap_keyidx); | ||||
|                 precompiled_charsmap.assign(pc, pc + n_precompiled_charsmap); | ||||
| #ifdef IS_BIG_ENDIAN | ||||
|  |  | |||
|  | @ -0,0 +1,64 @@ | |||
| From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001 | ||||
| From: jmorganca <jmorganca@gmail.com> | ||||
| Date: Wed, 5 Mar 2025 17:41:07 -0800 | ||||
| Subject: [PATCH] fix string arr kv loading | ||||
| 
 | ||||
| ---
 | ||||
|  ggml/include/gguf.h | 1 + | ||||
|  ggml/src/gguf.cpp   | 7 +++++-- | ||||
|  src/llama-vocab.cpp | 2 +- | ||||
|  3 files changed, 7 insertions(+), 3 deletions(-) | ||||
| 
 | ||||
| diff --git a/ggml/include/gguf.h b/ggml/include/gguf.h
 | ||||
| index 79ee2020..3efb22f0 100644
 | ||||
| --- a/ggml/include/gguf.h
 | ||||
| +++ b/ggml/include/gguf.h
 | ||||
| @@ -114,6 +114,7 @@ extern "C" {
 | ||||
|      // get raw pointer to the first element of the array with the given key_id | ||||
|      // for bool arrays, note that they are always stored as int8 on all platforms (usually this makes no difference) | ||||
|      GGML_API const void * gguf_get_arr_data(const struct gguf_context * ctx, int64_t key_id); | ||||
| +    GGML_API size_t       gguf_get_arr_data_n(const struct gguf_context * ctx, int64_t key_id);
 | ||||
|   | ||||
|      // get ith C string from array with given key_id | ||||
|      GGML_API const char * gguf_get_arr_str (const struct gguf_context * ctx, int64_t key_id, size_t i); | ||||
| diff --git a/ggml/src/gguf.cpp b/ggml/src/gguf.cpp
 | ||||
| index ab13669c..f75b923f 100644
 | ||||
| --- a/ggml/src/gguf.cpp
 | ||||
| +++ b/ggml/src/gguf.cpp
 | ||||
| @@ -777,10 +777,14 @@ enum gguf_type gguf_get_arr_type(const struct gguf_context * ctx, int64_t key_id
 | ||||
|   | ||||
|  const void * gguf_get_arr_data(const struct gguf_context * ctx, int64_t key_id) { | ||||
|      GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
| -    GGML_ASSERT(ctx->kv[key_id].get_type() != GGUF_TYPE_STRING);
 | ||||
|      return ctx->kv[key_id].data.data(); | ||||
|  } | ||||
|   | ||||
| +size_t gguf_get_arr_data_n(const struct gguf_context * ctx, int64_t key_id) {
 | ||||
| +    GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx));
 | ||||
| +    return ctx->kv[key_id].data.size();
 | ||||
| +}
 | ||||
| +
 | ||||
|  const char * gguf_get_arr_str(const struct gguf_context * ctx, int64_t key_id, size_t i) { | ||||
|      GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|      GGML_ASSERT(ctx->kv[key_id].get_type() == GGUF_TYPE_STRING); | ||||
| @@ -874,7 +878,6 @@ const char * gguf_get_val_str(const struct gguf_context * ctx, int64_t key_id) {
 | ||||
|  const void * gguf_get_val_data(const struct gguf_context * ctx, int64_t key_id) { | ||||
|      GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|      GGML_ASSERT(ctx->kv[key_id].get_ne() == 1); | ||||
| -    GGML_ASSERT(ctx->kv[key_id].get_type() != GGUF_TYPE_STRING);
 | ||||
|      return ctx->kv[key_id].data.data(); | ||||
|  } | ||||
|   | ||||
| diff --git a/src/llama-vocab.cpp b/src/llama-vocab.cpp
 | ||||
| index c7ff28be..7a185443 100644
 | ||||
| --- a/src/llama-vocab.cpp
 | ||||
| +++ b/src/llama-vocab.cpp
 | ||||
| @@ -1443,7 +1443,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) {
 | ||||
|   | ||||
|              const int precompiled_charsmap_keyidx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_PRECOMPILED_CHARSMAP).c_str()); | ||||
|              if (precompiled_charsmap_keyidx != -1) { | ||||
| -                size_t n_precompiled_charsmap = gguf_get_arr_n(ctx, precompiled_charsmap_keyidx);
 | ||||
| +                size_t n_precompiled_charsmap = gguf_get_arr_data_n(ctx, precompiled_charsmap_keyidx);
 | ||||
|                  const char * pc = (const char *) gguf_get_arr_data(ctx, precompiled_charsmap_keyidx); | ||||
|                  precompiled_charsmap.assign(pc, pc + n_precompiled_charsmap); | ||||
|  #ifdef IS_BIG_ENDIAN | ||||
|  | @ -114,6 +114,7 @@ extern "C" { | |||
|     // get raw pointer to the first element of the array with the given key_id
 | ||||
|     // for bool arrays, note that they are always stored as int8 on all platforms (usually this makes no difference)
 | ||||
|     GGML_API const void * gguf_get_arr_data(const struct gguf_context * ctx, int64_t key_id); | ||||
|     GGML_API size_t       gguf_get_arr_data_n(const struct gguf_context * ctx, int64_t key_id); | ||||
| 
 | ||||
|     // get ith C string from array with given key_id
 | ||||
|     GGML_API const char * gguf_get_arr_str (const struct gguf_context * ctx, int64_t key_id, size_t i); | ||||
|  |  | |||
|  | @ -777,10 +777,14 @@ enum gguf_type gguf_get_arr_type(const struct gguf_context * ctx, int64_t key_id | |||
| 
 | ||||
| const void * gguf_get_arr_data(const struct gguf_context * ctx, int64_t key_id) { | ||||
|     GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|     GGML_ASSERT(ctx->kv[key_id].get_type() != GGUF_TYPE_STRING); | ||||
|     return ctx->kv[key_id].data.data(); | ||||
| } | ||||
| 
 | ||||
| size_t gguf_get_arr_data_n(const struct gguf_context * ctx, int64_t key_id) { | ||||
|     GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|     return ctx->kv[key_id].data.size(); | ||||
| } | ||||
| 
 | ||||
| const char * gguf_get_arr_str(const struct gguf_context * ctx, int64_t key_id, size_t i) { | ||||
|     GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|     GGML_ASSERT(ctx->kv[key_id].get_type() == GGUF_TYPE_STRING); | ||||
|  | @ -874,7 +878,6 @@ const char * gguf_get_val_str(const struct gguf_context * ctx, int64_t key_id) { | |||
| const void * gguf_get_val_data(const struct gguf_context * ctx, int64_t key_id) { | ||||
|     GGML_ASSERT(key_id >= 0 && key_id < gguf_get_n_kv(ctx)); | ||||
|     GGML_ASSERT(ctx->kv[key_id].get_ne() == 1); | ||||
|     GGML_ASSERT(ctx->kv[key_id].get_type() != GGUF_TYPE_STRING); | ||||
|     return ctx->kv[key_id].data.data(); | ||||
| } | ||||
| 
 | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue