[Metax] optimize flash attention backend (#5876)

This commit is contained in:
Neil Zhu
2026-01-06 09:52:09 +08:00
committed by GitHub
parent ab553b3b8b
commit 272a371635
6 changed files with 990 additions and 618 deletions
+1 -1
View File
@@ -645,9 +645,9 @@ elif paddle.device.is_compiled_with_custom_device("metax_gpu"):
"metax_ops/moe_ffn.cu",
"metax_ops/moe_reduce.cu",
"metax_ops/fused_moe.cu",
"metax_ops/apply_rope_qkv.cu",
"metax_ops/cache_kv_with_rope.cu",
"metax_ops/cpp_extensions.cc",
"metax_ops/split_merge_qkv.cu",
]
sources += find_end_files("gpu_ops/speculate_decoding", ".cu")