From 20649b9730e9234b3fb1947e987a1122d43d7824 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Wed, 4 Jun 2025 11:11:23 -0700 Subject: [PATCH] Small changes --- modules/llama_cpp_server.py | 1 - server.py | 11 ++++++----- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index 30823625..aa712541 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -398,7 +398,6 @@ class LlamaServer: def stop(self): """Stop the server process.""" if self.process: - logger.info("Terminating llama-server...") self.process.terminate() try: self.process.wait(timeout=5) diff --git a/server.py b/server.py index 516253f7..99d2e171 100644 --- a/server.py +++ b/server.py @@ -61,11 +61,12 @@ from modules.utils import gradio def signal_handler(sig, frame): logger.info("Received Ctrl+C. Shutting down Text generation web UI gracefully.") - # Try to stop the model if loaded - try: - shared.model.stop() - except: - pass + # Explicitly stop LlamaServer to avoid __del__ cleanup issues during shutdown + if shared.model and shared.model.__class__.__name__ == 'LlamaServer': + try: + shared.model.stop() + except: + pass sys.exit(0)