mirror of https://github.com/google/gemma.cpp.git
* Adds and uses a new `AttentionActivationPtrs` that holds non-owning `MatPtrs`. Acts as a view into `AttentionActivations`. * Updates `QBatch` to hold non-owning `MatPtr`s to the kv caches. * Enables the `MatPtrT` default constructor for simpler initializations. * Pulls out and passes `LayerWeightsPtrs::query_norm_scale` directly. While `LayerWeightsPtrs` already held non-owning `MatPtr`s, this change avoids the need to find and construct several empty weight tensors just to construct one `query_norm_scale` tensor. PiperOrigin-RevId: 824584177 |
||
|---|---|---|
| .. | ||
| bench_matmul.cc | ||
| dot-inl.h | ||
| dot_test.cc | ||
| fp_arith-inl.h | ||
| matmul-inl.h | ||
| matmul.cc | ||
| matmul.h | ||
| matmul_static-inl.h | ||
| matmul_static.h | ||
| matmul_static_bf16.cc | ||
| matmul_static_f32.cc | ||
| matmul_static_i8.cc | ||
| matmul_static_nuq.cc | ||
| matmul_static_sfp.cc | ||
| matmul_test.cc | ||
| ops-inl.h | ||
| ops.h | ||
| ops_test.cc | ||
| sum-inl.h | ||