mirror of https://github.com/google/gemma.cpp.git
Move test placeholder to a later pos.
PiperOrigin-RevId: 644808456
This commit is contained in:
parent
0e612d9a20
commit
658fb3e506
|
|
@ -48,6 +48,7 @@
|
|||
#include "gemma/gemma.h"
|
||||
#include "gemma/weights.h"
|
||||
// Placeholder for internal test1, do not remove
|
||||
// Placeholder for internal test4, do not remove
|
||||
#include "hwy/aligned_allocator.h"
|
||||
#include "hwy/base.h"
|
||||
#include "hwy/contrib/thread_pool/thread_pool.h"
|
||||
|
|
@ -600,8 +601,6 @@ static void AddFromBatched(size_t num_tokens, const float* other, float* x,
|
|||
}
|
||||
}
|
||||
|
||||
// Placeholder for internal test3, do not remove
|
||||
|
||||
template <size_t kBatchSize, typename WeightArrayT, class TConfig>
|
||||
HWY_NOINLINE void EmbedToken(int token, size_t token_idx, size_t pos,
|
||||
const WeightArrayT& weights,
|
||||
|
|
@ -717,8 +716,6 @@ HWY_NOINLINE void Transformer(const int* tokens, size_t num_tokens, size_t pos,
|
|||
}
|
||||
}
|
||||
|
||||
// Placeholder for internal test4, do not remove
|
||||
|
||||
RMSNormInplaceBatched<kBatchSize>(num_tokens, weights.final_norm_scale.data(),
|
||||
activations.x.data(), kModelDim);
|
||||
if (layers_output) {
|
||||
|
|
@ -771,6 +768,8 @@ Activations<TConfig, kBatchSize>& GetActivations(const ByteStorageT& state_u8) {
|
|||
|
||||
} // namespace
|
||||
|
||||
// Placeholder for internal test3, do not remove
|
||||
|
||||
template <class TConfig>
|
||||
void GenerateT(const ByteStorageT& weights_u8, const ByteStorageT& prefill_u8,
|
||||
const ByteStorageT& decode_u8,
|
||||
|
|
|
|||
Loading…
Reference in New Issue