diff options
Diffstat (limited to 'candle-kernels/src/reduce.cu')
-rw-r--r-- | candle-kernels/src/reduce.cu | 14 |
1 files changed, 6 insertions, 8 deletions
diff --git a/candle-kernels/src/reduce.cu b/candle-kernels/src/reduce.cu index aaac24a1..079c3708 100644 --- a/candle-kernels/src/reduce.cu +++ b/candle-kernels/src/reduce.cu @@ -70,10 +70,9 @@ static __device__ __forceinline__ float warp_reduce_sum(float x) { // LayerNorm implementation adapted from ggml, accumulation is made using f32. // https://github.com/ggerganov/llama.cpp/blob/d59bd97065cd7ded6c4ecab54b1d5e0b1b11e318/ggml-cuda.cu#L477 template <typename T> -__device__ void layernorm(const T * x, T * dst, const T * alpha, const T * beta, const int ncols, const float eps) { +__device__ void layernorm(const T * x, T * dst, const T * alpha, const T * beta, const int ncols, const int block_size, const float eps) { const int row = blockIdx.x*blockDim.y + threadIdx.y; const int tid = threadIdx.x; - const int block_size = blockDim.x; float2 mean_var = make_float2(0.f, 0.f); @@ -134,10 +133,9 @@ __device__ void layernorm(const T * x, T * dst, const T * alpha, const T * beta, // RmsNorm implementation adapted from ggml, accumulation is made using f32. // https://github.com/ggerganov/llama.cpp/blob/d59bd97065cd7ded6c4ecab54b1d5e0b1b11e318/ggml-cuda.cu#L523 template <typename T> -__device__ void rmsnorm(const T * x, T * dst, const T * alpha, const int ncols, const float eps) { +__device__ void rmsnorm(const T * x, T * dst, const T * alpha, const int ncols, const int block_size, const float eps) { const int row = blockIdx.x*blockDim.y + threadIdx.y; const int tid = threadIdx.x; - const int block_size = blockDim.x; float tmp = 0.0f; // partial sum for thread in warp @@ -530,15 +528,15 @@ fast_argmax(const size_t src_numel, const size_t el_to_sum_per_block, #define RMSNORM_OP(TYPENAME, FN_NAME) \ extern "C" __global__ void FN_NAME( \ const TYPENAME *src, TYPENAME *dst, const TYPENAME *alpha, \ - const int n_cols, const float eps) { \ - rmsnorm<TYPENAME>(src, dst, alpha, n_cols, eps); \ + const int n_cols, const int block_size, const float eps) { \ + rmsnorm<TYPENAME>(src, dst, alpha, n_cols, block_size, eps); \ } \ #define LAYERNORM_OP(TYPENAME, FN_NAME) \ extern "C" __global__ void FN_NAME( \ const TYPENAME *src, TYPENAME *dst, const TYPENAME *alpha, \ - const TYPENAME *beta, const int n_cols, const float eps) { \ - layernorm<TYPENAME>(src, dst, alpha, beta, n_cols, eps); \ + const TYPENAME *beta, const int n_cols, const int block_size, const float eps) { \ + layernorm<TYPENAME>(src, dst, alpha, beta, n_cols, block_size, eps); \ } \ #define ROPE_OP(TYPENAME, FN_NAME, FN_NAME_I, FN_NAME_THD) \ |