115 lines
4.2 KiB
Python
115 lines
4.2 KiB
Python
from __future__ import annotations
|
|
|
|
from fastapi import APIRouter, Depends, HTTPException, Request, status
|
|
from sqlalchemy.ext.asyncio import AsyncSession
|
|
|
|
from app.core.database.session import get_session
|
|
from app.models import User
|
|
from app.schemas.pipeline_chat_sch import PipelineGenerateRequest, PipelineGenerateResponse
|
|
from app.services.pipeline_dialog_service import DialogAccessError, PipelineDialogService
|
|
from app.services.pipeline_service import PipelineService
|
|
from app.utils.business_logger import log_business_event
|
|
from app.utils.token_manager import get_current_user
|
|
|
|
|
|
router = APIRouter(tags=["Pipelines"])
|
|
|
|
|
|
@router.post("/generate", response_model=PipelineGenerateResponse)
|
|
async def generate_pipeline(
|
|
payload: PipelineGenerateRequest,
|
|
request: Request,
|
|
session: AsyncSession = Depends(get_session),
|
|
current_user: User = Depends(get_current_user),
|
|
):
|
|
trace_id = getattr(request.state, "traceId", None)
|
|
log_business_event(
|
|
"pipeline_prompt_received",
|
|
trace_id=trace_id,
|
|
user_id=str(current_user.id),
|
|
dialog_id=str(payload.dialog_id),
|
|
message_len=len(payload.message),
|
|
capability_ids_count=len(payload.capability_ids or []),
|
|
)
|
|
|
|
service = PipelineService(session)
|
|
dialog_service = PipelineDialogService(session)
|
|
try:
|
|
await dialog_service.append_user_message(
|
|
dialog_id=payload.dialog_id,
|
|
user_id=current_user.id,
|
|
content=payload.message,
|
|
)
|
|
dialog = await dialog_service.get_dialog(
|
|
dialog_id=payload.dialog_id,
|
|
user_id=current_user.id,
|
|
)
|
|
except DialogAccessError as exc:
|
|
detail = str(exc)
|
|
log_business_event(
|
|
"pipeline_prompt_rejected",
|
|
trace_id=trace_id,
|
|
user_id=str(current_user.id),
|
|
dialog_id=str(payload.dialog_id),
|
|
reason=detail,
|
|
)
|
|
if "denied" in detail:
|
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail=detail) from exc
|
|
raise HTTPException(status_code=status.HTTP_404_NOT_FOUND, detail=detail) from exc
|
|
|
|
try:
|
|
result = await service.generate(
|
|
dialog_id=payload.dialog_id,
|
|
message=payload.message,
|
|
user_id=current_user.id,
|
|
capability_ids=payload.capability_ids,
|
|
previous_pipeline_id=dialog.last_pipeline_id,
|
|
)
|
|
except Exception as exc:
|
|
if "ollama" in str(exc).lower():
|
|
message_ru = "Не удалось обратиться к локальной модели Ollama. Проверьте OLLAMA_HOST/OLLAMA_MODEL и повторите запрос."
|
|
result = {
|
|
"status": "cannot_build",
|
|
"message_ru": message_ru,
|
|
"chat_reply_ru": message_ru,
|
|
"pipeline_id": None,
|
|
"nodes": [],
|
|
"edges": [],
|
|
"missing_requirements": ["ollama_unavailable"],
|
|
"context_summary": None,
|
|
}
|
|
else:
|
|
raise
|
|
|
|
response_payload = PipelineGenerateResponse(**result)
|
|
try:
|
|
await dialog_service.append_assistant_message(
|
|
dialog_id=payload.dialog_id,
|
|
user_id=current_user.id,
|
|
content=response_payload.chat_reply_ru or response_payload.message_ru,
|
|
assistant_payload=response_payload.model_dump(mode="json", exclude_none=True),
|
|
)
|
|
except DialogAccessError as exc:
|
|
detail = str(exc)
|
|
log_business_event(
|
|
"pipeline_prompt_rejected",
|
|
trace_id=trace_id,
|
|
user_id=str(current_user.id),
|
|
dialog_id=str(payload.dialog_id),
|
|
reason=detail,
|
|
)
|
|
if "denied" in detail:
|
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail=detail) from exc
|
|
raise HTTPException(status_code=status.HTTP_404_NOT_FOUND, detail=detail) from exc
|
|
|
|
log_business_event(
|
|
"pipeline_prompt_processed",
|
|
trace_id=trace_id,
|
|
user_id=str(current_user.id),
|
|
dialog_id=str(payload.dialog_id),
|
|
result_status=response_payload.status,
|
|
pipeline_id=str(response_payload.pipeline_id) if response_payload.pipeline_id else None,
|
|
)
|
|
|
|
return response_payload
|