Logo
Explore Help
Sign In
apps/FastDeploy
1
0
Fork 0
You've already forked FastDeploy
mirror of https://github.com/PaddlePaddle/FastDeploy.git synced 2026-04-23 17:11:21 +08:00
Code Issues Actions 10 Packages Projects Releases Wiki Activity
Files
3b0bdbae657c00cbe0a4fff098d23cc5e298ffdd
FastDeploy/fastdeploy/model_executor/layers/attention
T
History
Lucas 0a0c74e717 [XPU] Support PaddleOCR-VL model for XPU (#4529)
* [XPU] support PaddleOCR-VL in XPU

* [XPU] fix PaddleOCR-VL pos_emb_type
2025-10-28 20:35:04 +08:00
..
ops
Support GPT-OSS-BF16 (#4240)
2025-10-20 14:44:58 +08:00
__init__.py
【FIX】Change the name of sparse attn from moba to plas (#4006) (#4076)
2025-09-23 10:26:40 +08:00
append_attn_backend.py
[Speculative Decoding][MTP]Support mtp in epdptp mode (#4614)
2025-10-28 16:02:47 +08:00
attention_selecter.py
…
attention.py
Support GPT-OSS-BF16 (#4240)
2025-10-20 14:44:58 +08:00
base_attention_backend.py
…
block_multihead_attn_backend.py
[FDConfig]Remove max_model_len in FDConfig (#4350)
2025-10-11 14:04:17 +08:00
flash_attn_backend.py
Support GPT-OSS-BF16 (#4240)
2025-10-20 14:44:58 +08:00
iluvatar_attn_backend.py
[Iluvatar GPU] Adapt VL model (#4313)
2025-10-17 16:13:38 +08:00
mla_attention_backend.py
[Optimization] Fuse get_max_len and get_kv_max_len (#4369)
2025-10-13 20:35:00 +08:00
moba_attention_backend.py
[FDConfig]Remove total_block_num/dtype/block_size/enc_dec_block_num in ParallelConfig (#4400)
2025-10-16 20:00:37 +08:00
native_paddle_backend.py
…
utils.py
…
xpu_attn_backend.py
[XPU] Support PaddleOCR-VL model for XPU (#4529)
2025-10-28 20:35:04 +08:00
Powered by Gitea Version: 1.26.0 Page: 1293ms Template: 85ms
Auto
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API