@@ -314,6 +314,49 @@ def decode(self, *args, **kwargs):
314314 "openai_force_temperature_one" : True ,
315315 },
316316
317+ "gpt-4.1" :{
318+ "fn_with_ui" : chatgpt_ui ,
319+ "fn_without_ui" : chatgpt_noui ,
320+ "has_multimodal_capacity" : True ,
321+ "endpoint" : openai_endpoint ,
322+ "max_token" : 828000 ,
323+ "tokenizer" : tokenizer_gpt4 ,
324+ "token_cnt" : get_token_num_gpt4 ,
325+ },
326+
327+ "gpt-4.1-mini" :{
328+ "fn_with_ui" : chatgpt_ui ,
329+ "fn_without_ui" : chatgpt_noui ,
330+ "has_multimodal_capacity" : True ,
331+ "endpoint" : openai_endpoint ,
332+ "max_token" : 828000 ,
333+ "tokenizer" : tokenizer_gpt4 ,
334+ "token_cnt" : get_token_num_gpt4 ,
335+ },
336+
337+ "o3" :{
338+ "fn_with_ui" : chatgpt_ui ,
339+ "fn_without_ui" : chatgpt_noui ,
340+ "has_multimodal_capacity" : True ,
341+ "endpoint" : openai_endpoint ,
342+ "max_token" : 828000 ,
343+ "tokenizer" : tokenizer_gpt4 ,
344+ "token_cnt" : get_token_num_gpt4 ,
345+ "openai_disable_system_prompt" : True ,
346+ "openai_disable_stream" : True ,
347+ "openai_force_temperature_one" : True ,
348+ },
349+
350+ "o4-mini" :{
351+ "fn_with_ui" : chatgpt_ui ,
352+ "fn_without_ui" : chatgpt_noui ,
353+ "has_multimodal_capacity" : True ,
354+ "endpoint" : openai_endpoint ,
355+ "max_token" : 828000 ,
356+ "tokenizer" : tokenizer_gpt4 ,
357+ "token_cnt" : get_token_num_gpt4 ,
358+ },
359+
317360 "gpt-4-turbo" : {
318361 "fn_with_ui" : chatgpt_ui ,
319362 "fn_without_ui" : chatgpt_noui ,
0 commit comments