[Models][Feature] Support new ERNIE reward model and add return_token_ids to reward API (#6638)

* reward model

* Add support for pooling-based inference in the reward model

* bugfix

---------

Co-authored-by: Jiang-Jia-Jun <163579578+Jiang-Jia-Jun@users.noreply.github.com>
This commit is contained in:
SunLei
2026-03-06 18:51:00 +08:00
committed by GitHub
parent caf73e8131
commit 5d9524fc3c
5 changed files with 22 additions and 6 deletions
@@ -1079,8 +1079,10 @@ PoolingChatRequest = EmbeddingChatRequest
class ChatRewardRequest(BaseModel):
model: Optional[str] = None
prompt_token_ids: Optional[List[int]] = None
messages: Union[List[Any], List[int]]
user: Optional[str] = None
return_token_ids: Optional[bool] = None
dimensions: Optional[int] = None
truncate_prompt_tokens: Optional[Annotated[int, Field(ge=-1)]] = None
@@ -1151,6 +1153,8 @@ class ChatRewardRequest(BaseModel):
class ChatRewardData(BaseModel):
index: Optional[int] = None
object: str = "reward"
prompt_token_ids: Optional[List[int]] = None
prompt_tokens: Optional[str] = None
score: List[float]