Skip to content

Commit 8a78963

Browse files
committedJan 10, 2025·
Fix rebase
1 parent 3c52f05 commit 8a78963

File tree

3 files changed

+11
-11
lines changed

3 files changed

+11
-11
lines changed
 

‎src/plugins/intel_gpu/src/graph/impls/ocl/paged_attention.cpp

+9-9
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
6969
void load(BinaryInputBuffer& ib) override {
7070
parent::load(ib);
7171
ib >> make_data(&has_scores_output, sizeof(bool));
72-
ib >> make_data(&has_rotation_coefficients, sizeof(bool));
72+
ib >> make_data(&has_rotated_blocks, sizeof(bool));
7373
if (is_dynamic()) {
7474
auto& kv_cache_update_kernel_selector = kv_cache_update_kernel_selector_t::Instance();
7575
auto kv_cache_update_kernel_impl = kv_cache_update_kernel_selector.GetImplementation(_kernels_data[Stage::KV_CACHE_UPDATE].kernelName);
@@ -83,7 +83,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
8383
auto pa_sdpa_kernel_impl = pa_sdpa_kernel_selector.GetImplementation(_kernels_data[Stage::PA_SDPA].kernelName);
8484
pa_sdpa_kernel_impl->GetUpdateDispatchDataFunc(_kernels_data[Stage::PA_SDPA]);
8585

86-
if (has_rotation_coefficients) {
86+
if (has_rotated_blocks) {
8787
auto& kv_cache_rotate_kernel_selector = kv_cache_rotate_kernel_selector_t::Instance();
8888
auto kv_cache_rotate_kernel_impl = kv_cache_rotate_kernel_selector.GetImplementation(_kernels_data[Stage::KV_CACHE_ROTATE].kernelName);
8989
kv_cache_rotate_kernel_impl->GetUpdateDispatchDataFunc(_kernels_data[Stage::KV_CACHE_ROTATE]);
@@ -94,7 +94,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
9494
void save(BinaryOutputBuffer& ob) const override {
9595
parent::save(ob);
9696
ob << make_data(&has_scores_output, sizeof(bool));
97-
ob << make_data(&has_rotation_coefficients, sizeof(bool));
97+
ob << make_data(&has_rotated_blocks, sizeof(bool));
9898
}
9999

100100
std::vector<layout> get_internal_buffer_layouts_impl() const override {
@@ -347,7 +347,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
347347

348348
std::vector<event::ptr> res_events;
349349
std::vector<event::ptr> dep_events = events;
350-
if (has_rotation_coefficients) {
350+
if (has_rotated_blocks) {
351351
execute_stage(dep_events, instance, res_events, Stage::KV_CACHE_ROTATE, is_mixed_mode);
352352
dep_events = res_events;
353353
}
@@ -472,7 +472,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
472472
config.has_const_scale_val = false;
473473
}
474474

475-
config.has_rotation_coefficients_input = desc->has_rotation_coefficients;
475+
config.has_rotated_blocks = desc->has_rotated_blocks;
476476

477477
if (desc->heads_num != desc->kv_heads_num) {
478478
config.broadcast_axis = 1;
@@ -752,7 +752,7 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
752752
for (const auto& input_layout : impl_param.input_layouts)
753753
input_tensors.emplace_back(convert_data_tensor(input_layout));
754754

755-
if (has_rotation_coefficients) {
755+
if (has_rotated_blocks) {
756756
auto kv_cache_rotate_kernel_params = get_kv_cache_rotate_kernel_params(impl_param, input_tensors, impl_param.is_dynamic());
757757
(_kernels_data[Stage::KV_CACHE_ROTATE].update_dispatch_data_func)(kv_cache_rotate_kernel_params, _kernels_data[Stage::KV_CACHE_ROTATE]);
758758
}
@@ -792,22 +792,22 @@ struct paged_attention_impl : multi_stage_primitive<paged_attention> {
792792
auto& pa_sdpa_kernel_selector = pa_sdpa_kernel_selector_t::Instance();
793793
kernels_data.push_back(pa_sdpa_kernel_selector.get_best_kernel(pa_sdpa_kernel_params));
794794

795-
if (desc->has_rotation_coefficients) {
795+
if (desc->has_rotated_blocks) {
796796
auto kv_cache_rotate_kernel_params = get_kv_cache_rotate_kernel_params(impl_param, input_tensors, impl_param.is_dynamic());
797797
auto& kv_cache_rotate_kernel_selector = kv_cache_rotate_kernel_selector_t::Instance();
798798
kernels_data.push_back(kv_cache_rotate_kernel_selector.get_best_kernel(kv_cache_rotate_kernel_params));
799799
}
800800

801801
auto impl = cldnn::make_unique<paged_attention_impl>(kernels_data);
802802
impl->has_scores_output = desc->has_scores_output();
803-
impl->has_rotation_coefficients = desc->has_rotation_coefficients;
803+
impl->has_rotated_blocks = desc->has_rotated_blocks;
804804

805805
return impl;
806806
}
807807

808808
private:
809809
bool has_scores_output = false;
810-
bool has_rotation_coefficients = false;
810+
bool has_rotated_blocks = false;
811811
};
812812

813813
namespace detail {

‎src/plugins/intel_gpu/src/plugin/ops/paged_attention.cpp

+1-1
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ static void CreatePagedAttentionExtensionOp(ProgramBuilder& p, const std::shared
6161
std::shared_ptr<ov::op::v0::Constant> alibi_const = std::dynamic_pointer_cast<ov::op::v0::Constant>(op->get_input_node_shared_ptr(alibi_idx));
6262
OPENVINO_ASSERT(alibi_const != nullptr);
6363
prim.has_alibi = ov::shape_size(alibi_const->get_output_shape(0)) > 0;
64-
prim.has_rotation_coefficients = op->get_input_size() == 16;
64+
prim.has_rotated_blocks = op->get_input_size() == 16;
6565

6666
prim.num_outputs = 1;
6767

‎src/plugins/intel_gpu/tests/unit/test_cases/paged_attention_gpu_test.cpp

+1-1
Original file line numberDiff line numberDiff line change
@@ -752,7 +752,7 @@ struct PagedAttentionTest : public ::testing::TestWithParam<T> {
752752
pa_prim.scale_val = pam.get_default_scale();
753753
pa_prim.has_alibi = false;
754754
pa_prim.num_outputs = p.scores_output ? 2 : 1;
755-
pa_prim.has_rotation_coefficients = p.rotation_config.apply_rotation;
755+
pa_prim.has_rotated_blocks = p.rotation_config.apply_rotation;
756756

757757
topology topology;
758758

0 commit comments

Comments
 (0)