@@ -106,7 +106,7 @@ def parse_args():
106106@app .exception_handler (RequestValidationError )
107107async def validation_exception_handler (_ , exc ):
108108 err = openai_serving_chat .create_error_response (message = str (exc ))
109- return JSONResponse (err .dict (), status_code = HTTPStatus .BAD_REQUEST )
109+ return JSONResponse (err .model_dump (), status_code = HTTPStatus .BAD_REQUEST )
110110
111111
112112@app .get ("/health" )
@@ -118,30 +118,36 @@ async def health() -> Response:
118118@app .get ("/v1/models" )
119119async def show_available_models ():
120120 models = await openai_serving_chat .show_available_models ()
121- return JSONResponse (content = models .dict ())
121+ return JSONResponse (content = models .model_dump ())
122122
123123
124124@app .post ("/v1/chat/completions" )
125125async def create_chat_completion (request : ChatCompletionRequest ,
126126 raw_request : Request ):
127127 generator = await openai_serving_chat .create_chat_completion (
128128 request , raw_request )
129- if request .stream and not isinstance (generator , ErrorResponse ):
129+ if isinstance (generator , ErrorResponse ):
130+ return JSONResponse (content = generator .model_dump (),
131+ status_code = generator .code )
132+ if request .stream :
130133 return StreamingResponse (content = generator ,
131134 media_type = "text/event-stream" )
132135 else :
133- return JSONResponse (content = generator .dict ())
136+ return JSONResponse (content = generator .model_dump ())
134137
135138
136139@app .post ("/v1/completions" )
137140async def create_completion (request : CompletionRequest , raw_request : Request ):
138141 generator = await openai_serving_completion .create_completion (
139142 request , raw_request )
140- if request .stream and not isinstance (generator , ErrorResponse ):
143+ if isinstance (generator , ErrorResponse ):
144+ return JSONResponse (content = generator .model_dump (),
145+ status_code = generator .code )
146+ if request .stream :
141147 return StreamingResponse (content = generator ,
142148 media_type = "text/event-stream" )
143149 else :
144- return JSONResponse (content = generator .dict ())
150+ return JSONResponse (content = generator .model_dump ())
145151
146152
147153if __name__ == "__main__" :
0 commit comments