mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2026-04-23 00:17:25 +08:00
support multi-step draft-model with cudagraph (#5886)
This commit is contained in:
@@ -2155,51 +2155,12 @@ class GPUModelRunner(ModelRunnerBase):
|
||||
),
|
||||
batch_size=int(capture_size / (self.speculative_config.num_speculative_tokens + 1)),
|
||||
in_capturing=True,
|
||||
expected_decode_len=self.speculative_config.num_speculative_tokens,
|
||||
expected_decode_len=self.speculative_config.num_speculative_tokens * 2 + 1,
|
||||
accept_all_drafts=True,
|
||||
)
|
||||
logger.info(
|
||||
f"Warm up the Target model with the num_tokens:{capture_size}, expected_decode_len:{self.speculative_config.num_speculative_tokens}"
|
||||
f"Warm up the model with the num_tokens:{capture_size}, expected_decode_len:{self.speculative_config.num_speculative_tokens}"
|
||||
)
|
||||
if self.graph_opt_config.draft_model_use_cudagraph:
|
||||
# Capture Draft Model without bsz 1
|
||||
# NOTE(liujundong): expected_decode_len = 1, will affect mtp capture in cudagraph
|
||||
for batch_size in sorted(capture_sizes, reverse=True):
|
||||
if batch_size == 1:
|
||||
logger.info("Skip token_num = 1, when capture Draft model for mtp")
|
||||
else:
|
||||
assert batch_size % 2 == 0
|
||||
self._dummy_run(
|
||||
num_tokens=(
|
||||
self.scheduler_config.max_num_seqs
|
||||
if self.scheduler_config.splitwise_role == "decode"
|
||||
else self.scheduler_config.max_num_batched_tokens
|
||||
),
|
||||
batch_size=int(batch_size / 2),
|
||||
in_capturing=True,
|
||||
expected_decode_len=3,
|
||||
accept_all_drafts=True,
|
||||
)
|
||||
logger.info(
|
||||
f"Warm up the Draft model with the num_tokens:{batch_size}, expected_decode_len:{3}"
|
||||
)
|
||||
# Capture Draft Model with bsz 1
|
||||
if 1 in capture_sizes:
|
||||
self._dummy_run(
|
||||
num_tokens=(
|
||||
self.scheduler_config.max_num_seqs
|
||||
if self.scheduler_config.splitwise_role == "decode"
|
||||
else self.scheduler_config.max_num_batched_tokens
|
||||
),
|
||||
batch_size=int(1),
|
||||
in_capturing=True,
|
||||
expected_decode_len=3,
|
||||
accept_all_drafts=False,
|
||||
reject_all_drafts=True,
|
||||
)
|
||||
logger.info(
|
||||
f"Warm up the Draft model with the num_tokens:{batch_size}, expected_decode_len:{3}"
|
||||
)
|
||||
else:
|
||||
for batch_size in sorted(capture_sizes, reverse=True):
|
||||
self._dummy_run(
|
||||
|
||||
Reference in New Issue
Block a user