diff --git a/src/llama-quant.cpp b/src/llama-quant.cpp index f2dab6a898..b8eb12690e 100644 --- a/src/llama-quant.cpp +++ b/src/llama-quant.cpp @@ -1088,14 +1088,12 @@ static std::unordered_map target_bpw_type( } } - auto side_data = [&](const std::unordered_map> * m, const std::string & tensor_name) -> std::pair { - if (!m) { return {nullptr, 0}; } + auto side_data = [&](const std::unordered_map> * m, const std::string & tensor_name) { + if (!m) { return std::pair{nullptr, 0}; } const std::string key = remap_imatrix(tensor_name, mapped); const auto it = m->find(key); - if (it == m->end()) { return {nullptr, 0}; } - - return { it->second.data(), it->second.size() }; + return it == m->end() ? std::pair{nullptr, 0} : std::pair{ it->second.data(), it->second.size() }; }; // Copy this row's side data (values and activations), or broadcasts to all slices @@ -1105,9 +1103,7 @@ static std::unordered_map target_bpw_type( const size_t want = (size_t)ne2 * (size_t)n_per_row; if (src_sz == want) { - dst.resize(want); - std::memcpy(dst.data(), src, want * sizeof(float)); - + dst.assign(src, src + want); return; } if (src_sz == (size_t)n_per_row) { @@ -1115,7 +1111,6 @@ static std::unordered_map target_bpw_type( for (int64_t s = 0; s < ne2; ++s) { std::memcpy(dst.data() + s * n_per_row, src, n_per_row * sizeof(float)); } - return; }