[Metax] optimize flash attention backend (#5876)

This commit is contained in:
Neil Zhu
2026-01-06 09:52:09 +08:00
committed by GitHub
parent ab553b3b8b
commit 272a371635
6 changed files with 990 additions and 618 deletions
@@ -59,7 +59,9 @@ elif current_platform.is_maca():
set_stop_value_multi_ends,
speculate_limit_thinking_content_length_v1,
speculate_limit_thinking_content_length_v2,
speculate_step_system_cache,
step_paddle,
step_system_cache,
update_inputs,
update_inputs_v1,
)