@@ -208,7 +208,6 @@ def process_request(self, request, max_model_len=None, **kwargs):
208208 str: error message
209209 """
210210 data_processor_logger .info (f"Start processing request: { request } " )
211- request .chat_template = kwargs .get ("chat_template" )
212211 request = self ._apply_default_parameters (request )
213212 if request .get ("eos_token_ids" ) is None or len (request .eos_token_ids ) == 0 :
214213 request .eos_token_ids = self .eos_token_ids
@@ -242,7 +241,7 @@ def process_request(self, request, max_model_len=None, **kwargs):
242241 if self .tokenizer .chat_template is None :
243242 raise ValueError ("This model does not support chat_template." )
244243 task = request .to_dict ()
245- chat_template_kwargs = kwargs .get ("chat_template_kwargs" )
244+ chat_template_kwargs = kwargs .get ("chat_template_kwargs" , {} )
246245 if chat_template_kwargs :
247246 if isinstance (chat_template_kwargs , dict ):
248247 for k , v in chat_template_kwargs .items ():
@@ -251,7 +250,7 @@ def process_request(self, request, max_model_len=None, **kwargs):
251250 else :
252251 raise ValueError ("Invalid input: chat_template_kwargs must be a dict" )
253252 task .setdefault ("enable_thinking" , True )
254- request .prompt_token_ids = self .messages2ids (task )
253+ request .prompt_token_ids = self .messages2ids (task , ** chat_template_kwargs )
255254 else :
256255 raise ValueError (f"The request should have `input_ids`, `text` or `messages`: { request } ." )
257256
@@ -316,7 +315,7 @@ def process_request_dict(self, request, max_model_len=None, **kwargs):
316315 elif request .get ("messages" ):
317316 if self .tokenizer .chat_template is None :
318317 raise ValueError ("This model does not support chat_template." )
319- chat_template_kwargs = request .get ("chat_template_kwargs" )
318+ chat_template_kwargs = request .get ("chat_template_kwargs" , {} )
320319 if chat_template_kwargs :
321320 if isinstance (chat_template_kwargs , dict ):
322321 for k , v in chat_template_kwargs .items ():
@@ -325,7 +324,7 @@ def process_request_dict(self, request, max_model_len=None, **kwargs):
325324 else :
326325 raise ValueError ("Invalid input: chat_template_kwargs must be a dict" )
327326 request .setdefault ("enable_thinking" , True )
328- request ["prompt_token_ids" ] = self .messages2ids (request )
327+ request ["prompt_token_ids" ] = self .messages2ids (request , ** chat_template_kwargs )
329328 else :
330329 raise ValueError (f"Request must contain 'prompt_token_ids', 'prompt', or 'messages': { request } " )
331330
@@ -530,7 +529,7 @@ def text2ids(self, text, max_model_len):
530529
531530 return tokens ["input_ids" ][0 ]
532531
533- def messages2ids (self , request ):
532+ def messages2ids (self , request , ** kwargs ):
534533 """
535534 Convert multi-turn messages into ID sequences.
536535
@@ -547,7 +546,7 @@ def messages2ids(self, request):
547546 split_special_tokens = False ,
548547 add_special_tokens = False ,
549548 return_tensors = "pd" ,
550- chat_template = request . get ( "chat_template" , None ) ,
549+ ** kwargs ,
551550 )
552551 request ["text_after_process" ] = spliced_message
553552 req_id = None
0 commit comments