Logo
Explore Help
Sign In
apps/FastDeploy
1
0
Fork 0
You've already forked FastDeploy
mirror of https://github.com/PaddlePaddle/FastDeploy.git synced 2026-04-23 00:17:25 +08:00
Code Issues Actions 21 Packages Projects Releases Wiki Activity
Files
223b2f5d86d2fbecfce94f44f9de2ec7dfaeadc3
FastDeploy/custom_ops
T
History
lzy 223b2f5d86 Support setting communication groups in custom_allreduce and the all-to-all\transpose fused operator during the decoding phase. (#5917)
2026-01-12 14:09:39 +08:00
..
cpu_ops
c++ code format (#4527)
2025-10-22 17:59:50 +08:00
gpu_ops
Support setting communication groups in custom_allreduce and the all-to-all\transpose fused operator during the decoding phase. (#5917)
2026-01-12 14:09:39 +08:00
iluvatar_ops
[Iluvatar] Support V1_KVCACHE_SCHEDULER and paddleocr-vl rope mode (#5555)
2025-12-18 02:14:25 -08:00
metax_ops
[Metax] adapt to gemm interface on different versions of maca (#5905)
2026-01-07 10:02:24 +08:00
third_party
[setup optimize]Support git submodule (#4033)
2025-09-11 17:41:16 +08:00
utils
[Bugfix] Increase the shape of w4afp8 gemm (#5957)
2026-01-09 14:11:17 +08:00
xpu_ops
[XPU] fix dp4 (#5946)
2026-01-09 20:36:53 +08:00
0001-DeepGEMM-95e81b3.patch
[OP]Remove extra H2D in DeepGemm (#5262)
2025-11-28 14:23:44 +08:00
MANIFEST.in
[LLM] First commit the llm deployment code
2025-06-09 19:20:15 +08:00
setup_ops_cpu.py
polish code with new pre-commit rule (#2923)
2025-07-19 23:19:27 +08:00
setup_ops.py
[Metax] adapt to gemm interface on different versions of maca (#5905)
2026-01-07 10:02:24 +08:00
Powered by Gitea Version: 1.26.0 Page: 224ms Template: 6ms
Auto
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API