llama.cpp/.github/ISSUE_TEMPLATE
Johannes Gäßler b1f3a6e5db
llama: automatically set parameters not set by the user in such a way that maximizes GPU utilization (#16653)
* llama: automatically fit args to free memory

llama-fit-params tool

* fix CI

* hints for bug reports, ensure no reallocation

* fix segfault with Vulkan

* add llama-fit-params to CI

* fix CI

* fix CI

* fix CI

* minor adjustments

* fix assignment of 1 dense layer

* fix logger not being reset on model load failure

* remove --n-gpu-layer hint on model load failure

* fix llama-fit-params verbosity

* fix edge case

* fix typo [no ci]
2025-12-15 09:24:59 +01:00
..
010-bug-compilation.yml ggml: initial IBM zDNN backend (#14975) 2025-08-15 21:11:22 +08:00
011-bug-results.yml llama: automatically set parameters not set by the user in such a way that maximizes GPU utilization (#16653) 2025-12-15 09:24:59 +01:00
019-bug-misc.yml github : add cmd line field to bug report (#11090) 2025-01-06 16:34:49 +01:00
020-enhancement.yml repo : update links to new url (#11886) 2025-02-15 16:40:57 +02:00
030-research.yml repo : update links to new url (#11886) 2025-02-15 16:40:57 +02:00
040-refactor.yml repo : update links to new url (#11886) 2025-02-15 16:40:57 +02:00
config.yml repo : update links to new url (#11886) 2025-02-15 16:40:57 +02:00