Remove ModelVariables (model_utils) from application & optimize Tenant

This commit is contained in:
Josako
2025-05-20 10:17:08 +02:00
parent 70de4c0328
commit d789e431ca
17 changed files with 83 additions and 206 deletions

View File

@@ -12,7 +12,8 @@ from common.utils.business_event_context import current_event
from eveai_chat_workers.specialists.specialist_typing import SpecialistArguments, SpecialistResult
from eveai_chat_workers.chat_session_cache import get_chat_history
from common.models.interaction import Specialist
from common.utils.model_utils import get_model_variables, create_language_template, replace_variable_in_template
from common.utils.model_utils import create_language_template, replace_variable_in_template, \
get_template
from eveai_chat_workers.specialists.base_specialist import BaseSpecialistExecutor
from eveai_chat_workers.retrievers.retriever_typing import RetrieverArguments
@@ -38,9 +39,6 @@ class SpecialistExecutor(BaseSpecialistExecutor):
# Initialize retrievers
self.retrievers = self._initialize_retrievers()
# Initialize model variables
self.model_variables = get_model_variables(tenant_id)
@property
def type(self) -> str:
return "STANDARD_RAG_SPECIALIST"
@@ -68,8 +66,7 @@ class SpecialistExecutor(BaseSpecialistExecutor):
])
# Get LLM and template
llm = self.model_variables.get_llm(temperature=0.3)
template = self.model_variables.get_template('history')
template, llm = get_template("history", temperature=0.3)
language_template = create_language_template(template, language)
# Create prompt
@@ -179,11 +176,7 @@ class SpecialistExecutor(BaseSpecialistExecutor):
with current_event.create_span("Specialist RAG invocation"):
try:
self.update_progress(self.task_id, "EveAI Chain Start", {})
# Get LLM with specified temperature
llm = self.model_variables.get_llm(temperature=self.temperature)
# Get template
template = self.model_variables.get_template('rag')
template, llm = get_template("rag", self.temperature)
language_template = create_language_template(template, language)
full_template = replace_variable_in_template(
language_template,

View File

@@ -8,7 +8,7 @@ from langchain_core.prompts import ChatPromptTemplate
from common.models.interaction import Specialist
from common.utils.business_event_context import current_event
from common.utils.model_utils import get_model_variables, get_crewai_llm, create_language_template
from common.utils.model_utils import get_crewai_llm, create_language_template, get_template
from eveai_chat_workers.retrievers.retriever_typing import RetrieverArguments
from eveai_chat_workers.specialists.crewai_base_classes import EveAICrewAIAgent, EveAICrewAITask
from crewai.tools import BaseTool
@@ -41,9 +41,6 @@ class CrewAIBaseSpecialistExecutor(BaseSpecialistExecutor):
# Initialize retrievers
self.retrievers = self._initialize_retrievers()
# Initialize model variables
self.model_variables = get_model_variables(tenant_id)
# initialize the Flow
self.flow = None
@@ -212,8 +209,7 @@ class CrewAIBaseSpecialistExecutor(BaseSpecialistExecutor):
try:
with current_event.create_span("Specialist Detail Question"):
# Get LLM and template
llm = self.model_variables.get_llm(temperature=0.3)
template = cache_manager.prompts_config_cache.get_config('history').get('content', '')
template, llm = get_template("history", temperature=0.3)
language_template = create_language_template(template, language)
# Create prompt