This commit is contained in:
oobabooga 2025-05-15 21:19:19 -07:00
parent 8cb73b78e1
commit fd61297933
4 changed files with 12 additions and 11 deletions

View file

@ -1,14 +1,14 @@
import copy import copy
import time
import json import json
import time
from collections import deque from collections import deque
import tiktoken import tiktoken
from pydantic import ValidationError
from extensions.openai.errors import InvalidRequestError from extensions.openai.errors import InvalidRequestError
from extensions.openai.utils import debug_msg, getToolCallId, parseToolCall
from extensions.openai.typing import ToolDefinition from extensions.openai.typing import ToolDefinition
from pydantic import ValidationError from extensions.openai.utils import debug_msg, getToolCallId, parseToolCall
from modules import shared from modules import shared
from modules.chat import ( from modules.chat import (
generate_chat_prompt, generate_chat_prompt,

View file

@ -1,9 +1,9 @@
import base64 import base64
import os
import time
import json import json
import os
import random import random
import re import re
import time
import traceback import traceback
from typing import Callable, Optional from typing import Callable, Optional

View file

@ -1,10 +1,11 @@
import math import math
import random import random
import threading import threading
import torch
import chromadb import chromadb
import numpy as np import numpy as np
import posthog import posthog
import torch
from chromadb.config import Settings from chromadb.config import Settings
from chromadb.utils import embedding_functions from chromadb.utils import embedding_functions

View file

@ -1,15 +1,15 @@
from pathlib import Path from pathlib import Path
import torch
import tensorrt_llm import tensorrt_llm
import torch
from tensorrt_llm.runtime import ModelRunner, ModelRunnerCpp
from modules import shared from modules import shared
from modules.logging_colors import logger from modules.logging_colors import logger
from modules.text_generation import ( from modules.text_generation import (
get_max_prompt_length, get_max_prompt_length,
get_reply_from_output_ids get_reply_from_output_ids
) )
from tensorrt_llm.runtime import ModelRunner, ModelRunnerCpp
class TensorRTLLMModel: class TensorRTLLMModel: