sglang_v0.5.2/flashinfer_0.3.1/include/flashinfer/attention/default_decode_params.cuh

280 lines
8.5 KiB
Plaintext

/*
* Copyright (c) 2024 by FlashInfer team.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef FLASHINFER_DECODE_PARAMS_CUH_
#define FLASHINFER_DECODE_PARAMS_CUH_
#include <cuda_runtime.h>
#include <cstdint>
#include "../layout.cuh"
#include "../page.cuh"
namespace flashinfer {
template <typename DTypeQ_, typename DTypeKV_, typename DTypeO_>
struct SingleDecodeParams {
using DTypeQ = DTypeQ_;
using DTypeKV = DTypeKV_;
using DTypeO = DTypeO_;
using IdType = int32_t;
DTypeQ* q;
DTypeKV* k;
DTypeKV* v;
DTypeO* o;
float* lse;
float* maybe_alibi_slopes;
uint32_t kv_len;
uint32_t num_qo_heads;
uint32_t num_kv_heads;
uint32_t q_stride_n;
uint32_t q_stride_h;
uint32_t kv_stride_n;
uint32_t kv_stride_h;
int32_t window_left;
float logits_soft_cap;
float sm_scale;
float rope_rcp_scale;
float rope_rcp_theta;
uint32_t kv_chunk_size;
__device__ __host__ SingleDecodeParams()
: q(nullptr),
k(nullptr),
v(nullptr),
o(nullptr),
lse(nullptr),
maybe_alibi_slopes(nullptr),
kv_len(0),
num_qo_heads(0),
num_kv_heads(0),
q_stride_n(0),
q_stride_h(0),
kv_stride_n(0),
kv_stride_h(0),
window_left(0),
logits_soft_cap(0.0f),
sm_scale(0.0f),
rope_rcp_scale(0.0f),
rope_rcp_theta(0.0f),
kv_chunk_size(0) {}
__device__ __host__ SingleDecodeParams(DTypeQ* q, DTypeKV* k, DTypeKV* v, DTypeO* o,
float* maybe_alibi_slopes, uint32_t seq_len,
uint32_t num_qo_heads, uint32_t num_kv_heads,
QKVLayout kv_layout, uint32_t head_dim,
int32_t window_left, float logits_soft_cap, float sm_scale,
float rope_scale, float rope_theta)
: q(q),
k(k),
v(v),
o(o),
lse(nullptr),
maybe_alibi_slopes(maybe_alibi_slopes),
kv_len(seq_len),
num_qo_heads(num_qo_heads),
num_kv_heads(num_kv_heads),
q_stride_n(num_qo_heads * head_dim),
q_stride_h(head_dim),
kv_stride_n((kv_layout == QKVLayout::kNHD) ? num_kv_heads * head_dim : head_dim),
kv_stride_h((kv_layout == QKVLayout::kNHD) ? head_dim : seq_len * head_dim),
window_left(window_left),
logits_soft_cap(logits_soft_cap),
sm_scale(sm_scale),
rope_rcp_scale(1.f / rope_scale),
rope_rcp_theta(1.f / rope_theta),
kv_chunk_size(0) {}
__host__ __device__ __forceinline__ uint32_t get_qo_len(uint32_t batch_idx) const { return 1; }
__host__ __device__ __forceinline__ uint32_t get_kv_len(uint32_t batch_idx) const {
return kv_len;
}
};
template <typename DTypeQ_, typename DTypeKV_, typename DTypeO_, typename IdType_>
struct BatchDecodeParams {
using DTypeQ = DTypeQ_;
using DTypeKV = DTypeKV_;
using DTypeO = DTypeO_;
using IdType = IdType_;
DTypeQ* q;
IdType* q_rope_offset;
paged_kv_t<DTypeKV, IdType> paged_kv;
DTypeO* o;
float* lse;
float* maybe_alibi_slopes;
uint32_t padded_batch_size;
uint32_t num_qo_heads;
IdType q_stride_n;
IdType q_stride_h;
int32_t window_left;
float logits_soft_cap;
float sm_scale;
float rope_rcp_scale;
float rope_rcp_theta;
IdType* request_indices;
IdType* kv_tile_indices;
IdType* o_indptr;
IdType* kv_chunk_size_ptr;
bool* block_valid_mask;
bool partition_kv;
__device__ __host__ BatchDecodeParams()
: q(nullptr),
q_rope_offset(nullptr),
paged_kv(),
o(nullptr),
lse(nullptr),
maybe_alibi_slopes(nullptr),
padded_batch_size(0),
num_qo_heads(0),
q_stride_n(0),
q_stride_h(0),
window_left(0),
logits_soft_cap(0.0f),
sm_scale(0.0f),
rope_rcp_scale(0.0f),
rope_rcp_theta(0.0f),
request_indices(nullptr),
kv_tile_indices(nullptr),
o_indptr(nullptr),
kv_chunk_size_ptr(nullptr),
block_valid_mask(nullptr),
partition_kv(false) {}
__device__ __host__ BatchDecodeParams(DTypeQ* q, IdType* q_rope_offset,
paged_kv_t<DTypeKV, IdType> paged_kv, DTypeO* o, float* lse,
float* maybe_alibi_slopes, uint32_t num_qo_heads,
IdType q_stride_n, IdType q_stride_h, int32_t window_left,
float logits_soft_cap, float sm_scale, float rope_scale,
float rope_theta)
: q(q),
q_rope_offset(q_rope_offset),
paged_kv(paged_kv),
o(o),
lse(lse),
maybe_alibi_slopes(maybe_alibi_slopes),
padded_batch_size(0),
num_qo_heads(num_qo_heads),
q_stride_n(q_stride_n),
q_stride_h(q_stride_h),
window_left(window_left),
logits_soft_cap(logits_soft_cap),
sm_scale(sm_scale),
rope_rcp_scale(1.f / rope_scale),
rope_rcp_theta(1.f / rope_theta),
request_indices(nullptr),
kv_tile_indices(nullptr),
o_indptr(nullptr),
kv_chunk_size_ptr(nullptr),
block_valid_mask(nullptr),
partition_kv(false) {}
__host__ __device__ __forceinline__ int32_t get_qo_len(int32_t batch_idx) const { return 1; }
__host__ __device__ __forceinline__ int32_t get_kv_len(int32_t batch_idx) const {
return paged_kv.get_length(batch_idx);
}
};
template <typename DTypeQ_, typename DTypeKV_, typename DTypeO_, typename IdType_>
struct BatchDecodeParamsMLA {
using DTypeQ = DTypeQ_;
using DTypeKV = DTypeKV_;
using DTypeO = DTypeO_;
using IdType = IdType_;
DTypeQ* q_nope;
DTypeQ* q_pe;
DTypeO* o;
float* lse;
float sm_scale;
IdType* q_rope_offset;
paged_kv_mla_t<DTypeKV, IdType> paged_kv;
uint32_t padded_batch_size;
uint32_t num_qo_heads;
int32_t window_left;
float logits_soft_cap;
float rope_rcp_scale;
float rope_rcp_theta;
IdType* request_indices;
IdType* kv_tile_indices;
IdType* o_indptr;
IdType* kv_chunk_size_ptr;
bool* block_valid_mask;
bool partition_kv;
__device__ __host__ BatchDecodeParamsMLA()
: q_nope(nullptr),
q_pe(nullptr),
o(nullptr),
lse(nullptr),
sm_scale(0.0f),
q_rope_offset(nullptr),
paged_kv(),
padded_batch_size(0),
num_qo_heads(0),
window_left(0),
logits_soft_cap(0.0f),
rope_rcp_scale(0.0f),
rope_rcp_theta(0.0f),
request_indices(nullptr),
kv_tile_indices(nullptr),
o_indptr(nullptr),
kv_chunk_size_ptr(nullptr),
block_valid_mask(nullptr),
partition_kv(false) {}
__device__ __host__ BatchDecodeParamsMLA(DTypeQ* q_nope, DTypeQ* q_pe, IdType* q_rope_offset,
paged_kv_mla_t<DTypeKV, IdType> paged_kv, DTypeO* o,
float* lse, uint32_t num_qo_heads, int32_t window_left,
float logits_soft_cap, float sm_scale, float rope_scale,
float rope_theta)
: q_nope(q_nope),
q_pe(q_pe),
o(o),
lse(lse),
sm_scale(sm_scale),
q_rope_offset(q_rope_offset),
paged_kv(paged_kv),
padded_batch_size(0),
num_qo_heads(num_qo_heads),
window_left(window_left),
logits_soft_cap(logits_soft_cap),
rope_rcp_scale(1.f / rope_scale),
rope_rcp_theta(1.f / rope_theta),
request_indices(nullptr),
kv_tile_indices(nullptr),
o_indptr(nullptr),
kv_chunk_size_ptr(nullptr),
block_valid_mask(nullptr),
partition_kv(false) {}
__host__ __device__ __forceinline__ int32_t get_qo_len(int32_t batch_idx) const { return 1; }
__host__ __device__ __forceinline__ int32_t get_kv_len(int32_t batch_idx) const {
return paged_kv.get_length(batch_idx);
}
};
} // namespace flashinfer
#endif // FLASHINFER_DECODE_PARAMS_CUH_