Removing trailing whitespaces
This commit is contained in:
parent
26cc75ffcc
commit
ceccfb9ee6
|
|
@ -160,7 +160,7 @@ struct llama_file::impl {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#elif defined(__linux__)
|
#elif defined(__linux__)
|
||||||
impl(const char * fname, const char * mode) : impl(fname, mode, false) {}
|
impl(const char * fname, const char * mode) : impl(fname, mode, false) {}
|
||||||
|
|
||||||
impl(const char * fname, const char * mode, bool uncached_read) {
|
impl(const char * fname, const char * mode, bool uncached_read) {
|
||||||
if (uncached_read) {
|
if (uncached_read) {
|
||||||
|
|
|
||||||
|
|
@ -502,7 +502,7 @@ llama_model_loader::llama_model_loader(
|
||||||
|
|
||||||
get_key(llm_kv(LLM_KV_GENERAL_ARCHITECTURE), arch_name, false);
|
get_key(llm_kv(LLM_KV_GENERAL_ARCHITECTURE), arch_name, false);
|
||||||
llm_kv = LLM_KV(llm_arch_from_string(arch_name));
|
llm_kv = LLM_KV(llm_arch_from_string(arch_name));
|
||||||
|
|
||||||
#if defined(__linux__)
|
#if defined(__linux__)
|
||||||
files.emplace_back(new llama_file(fname.c_str(), "rb", !use_mmap));
|
files.emplace_back(new llama_file(fname.c_str(), "rb", !use_mmap));
|
||||||
#else
|
#else
|
||||||
|
|
@ -1047,21 +1047,21 @@ bool llama_model_loader::load_all_data(
|
||||||
off_t aligned_offset = offset & ~(alignment - 1);
|
off_t aligned_offset = offset & ~(alignment - 1);
|
||||||
off_t offset_from_alignment = offset - aligned_offset;
|
off_t offset_from_alignment = offset - aligned_offset;
|
||||||
size_t bytes_to_read = (offset_from_alignment + size + alignment - 1) & ~(alignment - 1);
|
size_t bytes_to_read = (offset_from_alignment + size + alignment - 1) & ~(alignment - 1);
|
||||||
|
|
||||||
void * raw_buffer = nullptr;
|
void * raw_buffer = nullptr;
|
||||||
int ret = posix_memalign(&raw_buffer, alignment, bytes_to_read);
|
int ret = posix_memalign(&raw_buffer, alignment, bytes_to_read);
|
||||||
if (ret != 0) {
|
if (ret != 0) {
|
||||||
throw std::runtime_error(format("posix_memalign failed with error %d", ret));
|
throw std::runtime_error(format("posix_memalign failed with error %d", ret));
|
||||||
}
|
}
|
||||||
|
|
||||||
struct aligned_buffer_deleter {
|
struct aligned_buffer_deleter {
|
||||||
void operator()(void * p) const { free(p); }
|
void operator()(void * p) const { free(p); }
|
||||||
};
|
};
|
||||||
std::unique_ptr<void, aligned_buffer_deleter> buffer(raw_buffer);
|
std::unique_ptr<void, aligned_buffer_deleter> buffer(raw_buffer);
|
||||||
|
|
||||||
file->seek(aligned_offset, SEEK_SET);
|
file->seek(aligned_offset, SEEK_SET);
|
||||||
file->read_raw(buffer.get(), bytes_to_read);
|
file->read_raw(buffer.get(), bytes_to_read);
|
||||||
|
|
||||||
uintptr_t actual_data = reinterpret_cast<uintptr_t>(buffer.get()) + offset_from_alignment;
|
uintptr_t actual_data = reinterpret_cast<uintptr_t>(buffer.get()) + offset_from_alignment;
|
||||||
memcpy(dest, reinterpret_cast<void *>(actual_data), size);
|
memcpy(dest, reinterpret_cast<void *>(actual_data), size);
|
||||||
};
|
};
|
||||||
|
|
@ -1150,7 +1150,7 @@ bool llama_model_loader::load_all_data(
|
||||||
|
|
||||||
// Wait for previous upload to complete before reusing buffer
|
// Wait for previous upload to complete before reusing buffer
|
||||||
ggml_backend_event_synchronize(events[buffer_idx]);
|
ggml_backend_event_synchronize(events[buffer_idx]);
|
||||||
|
|
||||||
// Read aligned chunk from file
|
// Read aligned chunk from file
|
||||||
file->read_raw(reinterpret_cast<void *>(ptr_dest_aligned), read_size);
|
file->read_raw(reinterpret_cast<void *>(ptr_dest_aligned), read_size);
|
||||||
|
|
||||||
|
|
@ -1163,7 +1163,7 @@ bool llama_model_loader::load_all_data(
|
||||||
ptr_data += offset_from_alignment;
|
ptr_data += offset_from_alignment;
|
||||||
data_to_copy -= offset_from_alignment;
|
data_to_copy -= offset_from_alignment;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Trim alignment padding at end of last chunk
|
// Trim alignment padding at end of last chunk
|
||||||
if (aligned_offset + bytes_read + read_size > offset + n_size) {
|
if (aligned_offset + bytes_read + read_size > offset + n_size) {
|
||||||
data_to_copy -= (read_end - (offset + n_size));
|
data_to_copy -= (read_end - (offset + n_size));
|
||||||
|
|
@ -1204,9 +1204,9 @@ bool llama_model_loader::load_all_data(
|
||||||
read_aligned_chunk(file.get(), weight->offs, read_buf.data(), n_size, alignment);
|
read_aligned_chunk(file.get(), weight->offs, read_buf.data(), n_size, alignment);
|
||||||
#else
|
#else
|
||||||
file->seek(weight->offs, SEEK_SET);
|
file->seek(weight->offs, SEEK_SET);
|
||||||
file->read_raw(read_buf.data(), n_size);
|
file->read_raw(read_buf.data(), n_size);
|
||||||
#endif
|
#endif
|
||||||
ggml_backend_tensor_set(cur, read_buf.data(), 0, n_size);
|
ggml_backend_tensor_set(cur, read_buf.data(), 0, n_size);
|
||||||
if (check_tensors && !ggml_validate_row_data(cur->type, read_buf.data(), n_size)) {
|
if (check_tensors && !ggml_validate_row_data(cur->type, read_buf.data(), n_size)) {
|
||||||
throw std::runtime_error(format("tensor '%s' has invalid data", ggml_get_name(cur)));
|
throw std::runtime_error(format("tensor '%s' has invalid data", ggml_get_name(cur)));
|
||||||
}
|
}
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue