// Copyright 2024 Google LLC // SPDX-License-Identifier: Apache-2.0 // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // https://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // Command line tool to create compressed weights. // Compiles this file for multiple architectures via "foreach_target.h", to // which we pass the filename via macro 'argument'. #undef HWY_TARGET_INCLUDE #define HWY_TARGET_INCLUDE \ "compression/compress_weights.cc" // NOLINT #include "hwy/foreach_target.h" // IWYU pragma: keep // Must come after foreach_target.h to avoid redefinition errors. #include "compression/compress-inl.h" #include "hwy/highway.h" #ifndef GEMMA_COMPRESS_WEIGHTS_ONCE #define GEMMA_COMPRESS_WEIGHTS_ONCE #include #include #include // std::clamp #include #include #include #include // NOLINT #include "compression/io.h" // Path #include "gemma/common.h" // Model #include "gemma/weights.h" #include "gemma/weights_raw.h" #include "util/args.h" #include "hwy/base.h" #include "hwy/contrib/thread_pool/thread_pool.h" #include "hwy/profiler.h" namespace gcpp { // Setting this to true disables fread() calls that read the model file. constexpr bool kDryRunFread = false; namespace { float ScaleWeights(float* data, size_t len) { float maxabs = 0.0; for (size_t i = 0; i < len; ++i) { maxabs = std::max(maxabs, std::abs(data[i])); } const float kMaxRange = 1.875f; if (maxabs <= kMaxRange) { return 1.0f; } const float scale = maxabs / kMaxRange; const float inv_scale = 1.0f / scale; for (size_t i = 0; i < len; ++i) { data[i] *= inv_scale; } return scale; } #define READ_WEIGHTS(name) \ do { \ do_fread(&(layer_view->name), layer, #name, sizeof(layer_view->name)); \ } while (0) #define SCALE_WEIGHTS(name) \ do { \ if (ok && !kDryRunFread && scale_for_compression) { \ weights->scales[scale_pos++] = \ ScaleWeights(layer_view->name.data(), layer_view->name.size()); \ } \ } while (0) template struct LoadRawWeightsT { ByteStorageT operator()(const Path& checkpoint, hwy::ThreadPool& pool, bool scale_for_compression) const { PROFILER_ZONE("Startup.LoadWeights"); if (!checkpoint.Exists()) { HWY_ABORT("The model weights file '%s' does not exist.", checkpoint.path.c_str()); } ByteStorageT weights_u8 = AllocateWeightsF()(pool); auto* weights = reinterpret_cast*>(weights_u8.get()); size_t scale_pos = 0; FILE* fptr; if constexpr (kDryRunFread) { fprintf(stderr, "Dry-Run, not reading model-file.\n"); } else { fptr = fopen(checkpoint.path.c_str(), "rb"); if (fptr == nullptr) { HWY_ABORT("Failed to open model file %s - does it exist?", checkpoint.path.c_str()); } } bool ok = true; uint64_t total_size = 0; auto do_fread = [&](void* var, int layer, const char* name, size_t size) { if (layer == -1) { fprintf(stderr, "Loading Parameters (size %zu): %s\n", size, name); } else { fprintf(stderr, "Loading Parameters (layer=%d, size %zu): %s\n", layer, size, name); } if constexpr (!kDryRunFread) { ok &= 1 == fread(var, size, 1, fptr); total_size += size; } }; do_fread(&(weights->embedder_input_embedding), -1, "embedder_input_embedding", sizeof(weights->embedder_input_embedding)); do_fread(&(weights->final_norm_scale), -1, "final_norm_scale", sizeof(weights->final_norm_scale)); for (size_t layer = 0; layer < TConfig::kLayers; ++layer) { auto type = TConfig::kLayerConfig[layer]; LayerF* layer_view = weights->GetLayer(layer); // Make sure we don't have uninitialized memory. hwy::ZeroBytes(layer_view, sizeof(*layer_view)); if (type == LayerAttentionType::kGemma) { READ_WEIGHTS(attn_vec_einsum_w); READ_WEIGHTS(qkv_einsum_w); SCALE_WEIGHTS(attn_vec_einsum_w); SCALE_WEIGHTS(qkv_einsum_w); } else { READ_WEIGHTS(griffin.linear_x_w); READ_WEIGHTS(griffin.linear_x_biases); READ_WEIGHTS(griffin.linear_y_w); READ_WEIGHTS(griffin.linear_y_biases); READ_WEIGHTS(griffin.linear_out_w); READ_WEIGHTS(griffin.linear_out_biases); READ_WEIGHTS(griffin.conv_w); READ_WEIGHTS(griffin.conv_biases); READ_WEIGHTS(griffin.gate_w); READ_WEIGHTS(griffin.gate_biases); READ_WEIGHTS(griffin.a); SCALE_WEIGHTS(griffin.linear_x_w); SCALE_WEIGHTS(griffin.linear_y_w); SCALE_WEIGHTS(griffin.linear_out_w); SCALE_WEIGHTS(griffin.gate_w); } READ_WEIGHTS(gating_einsum_w); READ_WEIGHTS(linear_w); SCALE_WEIGHTS(gating_einsum_w); SCALE_WEIGHTS(linear_w); READ_WEIGHTS(pre_attention_norm_scale); READ_WEIGHTS(pre_ffw_norm_scale); if (TConfig::kPostNormScale) { READ_WEIGHTS(post_attention_norm_scale); READ_WEIGHTS(post_ffw_norm_scale); } if (TConfig::kFFBiases) { READ_WEIGHTS(ffw_gating_biases); READ_WEIGHTS(ffw_output_biases); } if (TConfig::kSoftmaxAttnOutputBiases && type == LayerAttentionType::kGemma) { READ_WEIGHTS(attention_output_biases); } } if (!ok) { HWY_ABORT( "Failed to read from %s - might be a directory, or too small? " "expected size: %d kB", checkpoint.path.c_str(), static_cast(total_size >> 10)); } if (!kDryRunFread) { HWY_ASSERT(0 == fclose(fptr)); if (scale_for_compression) { HWY_ASSERT(scale_pos == TConfig::kNumTensorScales); } } return weights_u8; } }; #undef READ_WEIGHTS #undef SCALE_WEIGHTS } // namespace ByteStorageT LoadRawWeights(const Path& weights, Model model_type, Type weight_type, hwy::ThreadPool& pool, bool scale_for_compression) { return CallForModelAndWeight( model_type, weight_type, weights, pool, scale_for_compression); } struct Args : public ArgsBase { static constexpr size_t kDefaultNumThreads = ~size_t{0}; void ChooseNumThreads() { if (num_threads == kDefaultNumThreads) { // This is a rough heuristic, replace with something better in the future. num_threads = static_cast(std::clamp( static_cast(std::thread::hardware_concurrency()) - 2, 1, 18)); } } public: Args(int argc, char* argv[]) { InitAndParse(argc, argv); ChooseNumThreads(); } // Returns error string or nullptr if OK. const char* Validate() { ModelTraining model_training; if (const char* err = ParseModelTypeAndTraining(model_type_str, model_type_, model_training)) { return err; } if (const char* err = ParseType(weight_type_str, weight_type_)) { return err; } if (weights.path.empty()) { return "Missing --weights flag, a file for the uncompressed model."; } if (compressed_weights.path.empty()) { return "Missing --compressed_weights flag, a file for the compressed " "model."; } if (!weights.Exists()) { return "Can't open file specified with --weights flag."; } return nullptr; } Path weights; // uncompressed weights file location Path compressed_weights; // compressed weights file location std::string model_type_str; std::string weight_type_str; size_t num_threads; template void ForEach(const Visitor& visitor) { visitor(weights, "weights", Path(), "Path to model weights (.bin) file.\n" " Required argument."); visitor(model_type_str, "model", std::string(), "Model type\n 2b-it = 2B parameters, instruction-tuned\n " "2b-pt = 2B parameters, pretrained\n 7b-it = 7B parameters " "instruction-tuned\n 7b-pt = 7B parameters, pretrained\n " "gr2b-it = griffin 2B parameters, instruction-tuned\n " "gr2b-pt = griffin 2B parameters, pretrained\n " " Required argument."); visitor(weight_type_str, "weight_type", std::string("sfp"), "Weight type\n f32 = float, bf16 = bfloat16, SFP = 8-bit FP\n" " Required argument."); visitor(compressed_weights, "compressed_weights", Path(), "Path name where compressed weights (.sbs) file will be written.\n" " Required argument."); visitor(num_threads, "num_threads", kDefaultNumThreads, // see ChooseNumThreads "Number of threads to use.\n Default = Estimate of the " "number of supported concurrent threads.", 2); } // Uninitialized before Validate, must call after that. gcpp::Model ModelType() const { return model_type_; } gcpp::Type WeightType() const { return weight_type_; } private: Model model_type_; Type weight_type_; }; void ShowHelp(gcpp::Args& args) { std::cerr << "Usage:\n./compress_weights --weights " " --model --compressed_weights \n"; std::cerr << "\n*Arguments*\n\n"; args.Help(); std::cerr << "\n"; } } // namespace gcpp #endif // GEMMA_COMPRESS_WEIGHTS_ONCE // SIMD code, compiled once per target. HWY_BEFORE_NAMESPACE(); namespace gcpp { namespace HWY_NAMESPACE { template void CompressWeights(const Path& weights_path, const Path& compressed_weights_path, Model model_type, Type weight_type, hwy::ThreadPool& pool) { if (!weights_path.Exists()) { HWY_ABORT("The model weights file '%s' does not exist.", weights_path.path.c_str()); } // Allocate compressed weights. using CWeights = CompressedWeights; ByteStorageT c_weights_u8 = AllocateSizeof(); CWeights* c_weights = reinterpret_cast(c_weights_u8.get()); new (&c_weights->c_layer_ptrs) CompressedLayerPointers(pool); // Get weights, compress, and store. const bool scale_for_compression = TConfig::kNumTensorScales > 0; const ByteStorageT weights_u8 = gcpp::LoadRawWeights( weights_path, model_type, weight_type, pool, scale_for_compression); WeightsF* weights = reinterpret_cast*>(weights_u8.get()); Compressor compressor(pool); ForEachTensor>(weights, *c_weights, compressor); compressor.AddScales(weights->scales.data(), weights->scales.size()); compressor.WriteAll(pool, compressed_weights_path); weights->layer_ptrs.~LayerPointers(); c_weights->c_layer_ptrs.~CompressedLayerPointers(); } } // namespace HWY_NAMESPACE } // namespace gcpp HWY_AFTER_NAMESPACE(); #if HWY_ONCE namespace gcpp { void Run(Args& args) { hwy::ThreadPool pool(args.num_threads); const Model model_type = args.ModelType(); const Type weight_type = args.WeightType(); GEMMA_EXPORT_AND_DISPATCH( model_type, weight_type, CompressWeights, (args.weights, args.compressed_weights, model_type, weight_type, pool)); } } // namespace gcpp int main(int argc, char** argv) { gcpp::Args args(argc, argv); if (gcpp::HasHelp(argc, argv)) { gcpp::ShowHelp(args); return 0; } if (const char* error = args.Validate()) { gcpp::ShowHelp(args); HWY_ABORT("\nInvalid args: %s", error); } gcpp::Run(args); return 0; } #endif // HWY_ONCE