diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index 30823625..aa712541 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -398,7 +398,6 @@ class LlamaServer: def stop(self): """Stop the server process.""" if self.process: - logger.info("Terminating llama-server...") self.process.terminate() try: self.process.wait(timeout=5) diff --git a/server.py b/server.py index 516253f7..99d2e171 100644 --- a/server.py +++ b/server.py @@ -61,11 +61,12 @@ from modules.utils import gradio def signal_handler(sig, frame): logger.info("Received Ctrl+C. Shutting down Text generation web UI gracefully.") - # Try to stop the model if loaded - try: - shared.model.stop() - except: - pass + # Explicitly stop LlamaServer to avoid __del__ cleanup issues during shutdown + if shared.model and shared.model.__class__.__name__ == 'LlamaServer': + try: + shared.model.stop() + except: + pass sys.exit(0)