Refer to --weights rather than --compressed_weights to simplify CLI docs

PiperOrigin-RevId: 634391135
This commit is contained in:
Paul Chang 2024-05-16 07:51:16 -07:00 committed by Copybara-Service
parent 8e641eb4cd
commit 82623bdc7f
2 changed files with 13 additions and 14 deletions

View File

@ -218,19 +218,18 @@ You can now run `gemma` from inside the `build/` directory.
`gemma` has the following required arguments: `gemma` has the following required arguments:
| Argument | Description | Example value | | Argument | Description | Example value |
| -------- | ----------- | ------------- | | ------------- | ---------------------------- | -------------------------- |
| `--model` | The model type. | `2b-it`, `2b-pt`, `7b-it`, `7b-pt`, ... (see above) | | `--model` | The model type. | `2b-it`, `2b-pt`, `7b-it`, `7b-pt`, ... (see above) |
| `--compressed_weights` | The compressed weights file. | `2b-it-sfp.sbs`, ... (see above) | | `--weights` | The compressed weights file. | `2b-it-sfp.sbs`, ... (see above) |
| `--tokenizer` | The tokenizer file. | `tokenizer.spm` | | `--tokenizer` | The tokenizer file. | `tokenizer.spm` |
`gemma` is invoked as: `gemma` is invoked as:
```sh ```sh
./gemma \ ./gemma \
--tokenizer [tokenizer file] \ --tokenizer [tokenizer file] \
--compressed_weights [compressed weights file] \ --weights [compressed weights file] \
--model [2b-it or 2b-pt or 7b-it or 7b-pt or ...] --model [2b-it or 2b-pt or 7b-it or 7b-pt or ...]
``` ```
@ -243,7 +242,7 @@ Example invocation for the following configuration:
```sh ```sh
./gemma \ ./gemma \
--tokenizer tokenizer.spm \ --tokenizer tokenizer.spm \
--compressed_weights 2b-it-sfp.sbs \ --weights 2b-it-sfp.sbs \
--model 2b-it --model 2b-it
``` ```
@ -262,8 +261,7 @@ tokenizer from the RecurrentGemma
[Kaggle](https://www.kaggle.com/models/google/recurrentgemma/gemmaCpp) as in [Kaggle](https://www.kaggle.com/models/google/recurrentgemma/gemmaCpp) as in
Step 1, and run the binary as follows: Step 1, and run the binary as follows:
`./gemma --tokenizer tokenizer.spm --model gr2b-it --compressed_weights 2b-it-sfp.sbs` `./gemma --tokenizer tokenizer.spm --model gr2b-it --weights 2b-it-sfp.sbs`
### Troubleshooting and FAQs ### Troubleshooting and FAQs
@ -382,7 +380,7 @@ For using the `gemma` executable as a command line tool, it may be useful to
create an alias for gemma.cpp with arguments fully specified: create an alias for gemma.cpp with arguments fully specified:
```sh ```sh
alias gemma2b="~/gemma.cpp/build/gemma -- --tokenizer ~/gemma.cpp/build/tokenizer.spm --compressed_weights ~/gemma.cpp/build/2b-it-sfp.sbs --model 2b-it --verbosity 0" alias gemma2b="~/gemma.cpp/build/gemma -- --tokenizer ~/gemma.cpp/build/tokenizer.spm --weights ~/gemma.cpp/build/2b-it-sfp.sbs --model 2b-it --verbosity 0"
``` ```
Replace the above paths with your own paths to the model and tokenizer paths Replace the above paths with your own paths to the model and tokenizer paths

View File

@ -79,11 +79,12 @@ void ShowHelp(gcpp::LoaderArgs& loader, gcpp::InferenceArgs& inference,
<< "\n\ngemma.cpp : a lightweight, standalone C++ inference engine\n" << "\n\ngemma.cpp : a lightweight, standalone C++ inference engine\n"
"==========================================================\n\n" "==========================================================\n\n"
"To run gemma.cpp, you need to " "To run gemma.cpp, you need to "
"specify 3 required model loading arguments:\n --tokenizer\n " "specify 3 required model loading arguments:\n"
"--compressed_weights\n" " --tokenizer\n"
" --weights\n"
" --model.\n"; " --model.\n";
std::cerr << "\n*Example Usage*\n\n./gemma --tokenizer tokenizer.spm " std::cerr << "\n*Example Usage*\n\n./gemma --tokenizer tokenizer.spm "
"--compressed_weights 2b-it-sfp.sbs --model 2b-it\n"; "--weights 2b-it-sfp.sbs --model 2b-it\n";
std::cerr << "\n*Model Loading Arguments*\n\n"; std::cerr << "\n*Model Loading Arguments*\n\n";
loader.Help(); loader.Help();
std::cerr << "\n*Inference Arguments*\n\n"; std::cerr << "\n*Inference Arguments*\n\n";