From 83bfd5c64b44e9eada63963e8aff05a608a7e90c Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sun, 18 May 2025 12:45:01 -0700 Subject: [PATCH] Fix API issues --- extensions/openai/script.py | 18 ++++++++++-------- 1 file changed, 10 insertions(+), 8 deletions(-) diff --git a/extensions/openai/script.py b/extensions/openai/script.py index b7394bc5..b6abae20 100644 --- a/extensions/openai/script.py +++ b/extensions/openai/script.py @@ -114,8 +114,8 @@ async def openai_completions(request: Request, request_data: CompletionRequest): if request_data.stream: async def generator(): - try: - async with streaming_semaphore: + async with streaming_semaphore: + try: response = OAIcompletions.stream_completions(to_dict(request_data), is_legacy=is_legacy) async for resp in iterate_in_threadpool(response): disconnected = await request.is_disconnected() @@ -123,8 +123,9 @@ async def openai_completions(request: Request, request_data: CompletionRequest): break yield {"data": json.dumps(resp)} - finally: - return + finally: + stop_everything_event() + return return EventSourceResponse(generator()) # SSE streaming @@ -145,8 +146,8 @@ async def openai_chat_completions(request: Request, request_data: ChatCompletion if request_data.stream: async def generator(): - try: - async with streaming_semaphore: + async with streaming_semaphore: + try: response = OAIcompletions.stream_chat_completions(to_dict(request_data), is_legacy=is_legacy) async for resp in iterate_in_threadpool(response): disconnected = await request.is_disconnected() @@ -154,8 +155,9 @@ async def openai_chat_completions(request: Request, request_data: ChatCompletion break yield {"data": json.dumps(resp)} - finally: - return + finally: + stop_everything_event() + return return EventSourceResponse(generator()) # SSE streaming