From 5efe017531050f942154146fea2935c7aff9ff3a Mon Sep 17 00:00:00 2001 From: ryan-mangeno Date: Mon, 15 Dec 2025 10:19:56 -0500 Subject: [PATCH] fixed whitespace and newline errors in editorconfig job --- src/llama-model.cpp | 4 ++-- src/llama-vocab.cpp | 2 +- src/models/modern-bert.cpp | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 0d3b2bb691..06fa1454a0 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -3166,7 +3166,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) { for(int i = 0; i < n_layer; ++i) { auto& layer = layers[i]; - + if ( i != 0 ) { layer.attn_norm = create_tensor(tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd}, 0); } else{ @@ -7803,7 +7803,7 @@ llama_rope_type llama_model_rope_type(const llama_model * model) { case LLM_ARCH_DBRX: case LLM_ARCH_BERT: case LLM_ARCH_JINA_BERT_V3: - case LLM_ARCH_MODERN_BERT: + case LLM_ARCH_MODERN_BERT: case LLM_ARCH_NOMIC_BERT: case LLM_ARCH_NOMIC_BERT_MOE: case LLM_ARCH_STABLELM: diff --git a/src/llama-vocab.cpp b/src/llama-vocab.cpp index 0c03ddcabd..ad8e1f9d72 100644 --- a/src/llama-vocab.cpp +++ b/src/llama-vocab.cpp @@ -1878,7 +1878,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) { tokenizer_pre == "jina-v2-es" || tokenizer_pre == "jina-v2-de" || tokenizer_pre == "a.x-4.0" || - tokenizer_pre == "mellum" || + tokenizer_pre == "mellum" || tokenizer_pre == "modern-bert" ) { pre_type = LLAMA_VOCAB_PRE_TYPE_GPT2; } else if ( diff --git a/src/models/modern-bert.cpp b/src/models/modern-bert.cpp index 90031c748b..bf158a9957 100644 --- a/src/models/modern-bert.cpp +++ b/src/models/modern-bert.cpp @@ -123,4 +123,4 @@ llm_build_modern_bert::llm_build_modern_bert(const llama_model & model, co // Explicit template instantiations template struct llm_build_modern_bert; -template struct llm_build_modern_bert; \ No newline at end of file +template struct llm_build_modern_bert;