mirror of https://github.com/google/gemma.cpp.git
Refer to --weights rather than --compressed_weights to simplify CLI docs
PiperOrigin-RevId: 634391135
This commit is contained in:
parent
8e641eb4cd
commit
82623bdc7f
14
README.md
14
README.md
|
|
@ -219,18 +219,17 @@ You can now run `gemma` from inside the `build/` directory.
|
||||||
`gemma` has the following required arguments:
|
`gemma` has the following required arguments:
|
||||||
|
|
||||||
| Argument | Description | Example value |
|
| Argument | Description | Example value |
|
||||||
| -------- | ----------- | ------------- |
|
| ------------- | ---------------------------- | -------------------------- |
|
||||||
| `--model` | The model type. | `2b-it`, `2b-pt`, `7b-it`, `7b-pt`, ... (see above) |
|
| `--model` | The model type. | `2b-it`, `2b-pt`, `7b-it`, `7b-pt`, ... (see above) |
|
||||||
| `--compressed_weights` | The compressed weights file. | `2b-it-sfp.sbs`, ... (see above) |
|
| `--weights` | The compressed weights file. | `2b-it-sfp.sbs`, ... (see above) |
|
||||||
| `--tokenizer` | The tokenizer file. | `tokenizer.spm` |
|
| `--tokenizer` | The tokenizer file. | `tokenizer.spm` |
|
||||||
|
|
||||||
|
|
||||||
`gemma` is invoked as:
|
`gemma` is invoked as:
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
./gemma \
|
./gemma \
|
||||||
--tokenizer [tokenizer file] \
|
--tokenizer [tokenizer file] \
|
||||||
--compressed_weights [compressed weights file] \
|
--weights [compressed weights file] \
|
||||||
--model [2b-it or 2b-pt or 7b-it or 7b-pt or ...]
|
--model [2b-it or 2b-pt or 7b-it or 7b-pt or ...]
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
@ -243,7 +242,7 @@ Example invocation for the following configuration:
|
||||||
```sh
|
```sh
|
||||||
./gemma \
|
./gemma \
|
||||||
--tokenizer tokenizer.spm \
|
--tokenizer tokenizer.spm \
|
||||||
--compressed_weights 2b-it-sfp.sbs \
|
--weights 2b-it-sfp.sbs \
|
||||||
--model 2b-it
|
--model 2b-it
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
@ -262,8 +261,7 @@ tokenizer from the RecurrentGemma
|
||||||
[Kaggle](https://www.kaggle.com/models/google/recurrentgemma/gemmaCpp) as in
|
[Kaggle](https://www.kaggle.com/models/google/recurrentgemma/gemmaCpp) as in
|
||||||
Step 1, and run the binary as follows:
|
Step 1, and run the binary as follows:
|
||||||
|
|
||||||
`./gemma --tokenizer tokenizer.spm --model gr2b-it --compressed_weights 2b-it-sfp.sbs`
|
`./gemma --tokenizer tokenizer.spm --model gr2b-it --weights 2b-it-sfp.sbs`
|
||||||
|
|
||||||
|
|
||||||
### Troubleshooting and FAQs
|
### Troubleshooting and FAQs
|
||||||
|
|
||||||
|
|
@ -382,7 +380,7 @@ For using the `gemma` executable as a command line tool, it may be useful to
|
||||||
create an alias for gemma.cpp with arguments fully specified:
|
create an alias for gemma.cpp with arguments fully specified:
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
alias gemma2b="~/gemma.cpp/build/gemma -- --tokenizer ~/gemma.cpp/build/tokenizer.spm --compressed_weights ~/gemma.cpp/build/2b-it-sfp.sbs --model 2b-it --verbosity 0"
|
alias gemma2b="~/gemma.cpp/build/gemma -- --tokenizer ~/gemma.cpp/build/tokenizer.spm --weights ~/gemma.cpp/build/2b-it-sfp.sbs --model 2b-it --verbosity 0"
|
||||||
```
|
```
|
||||||
|
|
||||||
Replace the above paths with your own paths to the model and tokenizer paths
|
Replace the above paths with your own paths to the model and tokenizer paths
|
||||||
|
|
|
||||||
|
|
@ -79,11 +79,12 @@ void ShowHelp(gcpp::LoaderArgs& loader, gcpp::InferenceArgs& inference,
|
||||||
<< "\n\ngemma.cpp : a lightweight, standalone C++ inference engine\n"
|
<< "\n\ngemma.cpp : a lightweight, standalone C++ inference engine\n"
|
||||||
"==========================================================\n\n"
|
"==========================================================\n\n"
|
||||||
"To run gemma.cpp, you need to "
|
"To run gemma.cpp, you need to "
|
||||||
"specify 3 required model loading arguments:\n --tokenizer\n "
|
"specify 3 required model loading arguments:\n"
|
||||||
"--compressed_weights\n"
|
" --tokenizer\n"
|
||||||
|
" --weights\n"
|
||||||
" --model.\n";
|
" --model.\n";
|
||||||
std::cerr << "\n*Example Usage*\n\n./gemma --tokenizer tokenizer.spm "
|
std::cerr << "\n*Example Usage*\n\n./gemma --tokenizer tokenizer.spm "
|
||||||
"--compressed_weights 2b-it-sfp.sbs --model 2b-it\n";
|
"--weights 2b-it-sfp.sbs --model 2b-it\n";
|
||||||
std::cerr << "\n*Model Loading Arguments*\n\n";
|
std::cerr << "\n*Model Loading Arguments*\n\n";
|
||||||
loader.Help();
|
loader.Help();
|
||||||
std::cerr << "\n*Inference Arguments*\n\n";
|
std::cerr << "\n*Inference Arguments*\n\n";
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue