[Optimization]update prompt & prompt_token_ids (#6334)

* fix prompt

* add unit test

* add unit test

* fix
This commit is contained in:
luukunn
2026-02-04 20:08:01 +08:00
committed by GitHub
parent bf78a48eb3
commit 765df94e6c
4 changed files with 28 additions and 31 deletions
+3 -10
View File
@@ -313,17 +313,10 @@ class DataProcessor(BaseDataProcessor):
# processing prompt_token_ids
if not request.get("prompt_token_ids"):
if request.get("prompt"):
prompt = request.get("prompt")
add_special_tokens = request.get("add_special_tokens", False)
assert isinstance(prompt, str) or (
isinstance(prompt, list) and all([isinstance(t, int) for t in prompt])
), f"prompt must be a string or a list of integers, but got {type(prompt)}"
if isinstance(prompt, list): # if prompt is a token id list
request["prompt_token_ids"] = prompt
else:
request["prompt_token_ids"] = self.text2ids(
request["prompt"], max_model_len, add_special_tokens=add_special_tokens
).tolist()
request["prompt_token_ids"] = self.text2ids(
request["prompt"], max_model_len, add_special_tokens=add_special_tokens
).tolist()
elif request.get("messages"):
if self.tokenizer.chat_template is None:
raise ValueError("This model does not support chat_template.")