feat: AI 서비스 및 AI 어시스턴트 전용 페이지 추가
- ai-service: Ollama 기반 AI 서비스 (분류, 시맨틱 검색, RAG Q&A, 패턴 분석) - AI 어시스턴트 페이지: 채팅형 Q&A, 시맨틱 검색, 패턴 분석, 분류 테스트 - 권한 시스템에 ai_assistant 페이지 등록 (기본 비활성) - 기존 페이지에 AI 기능 통합 (대시보드, 수신함, 관리함) - docker-compose, gateway, nginx 설정 업데이트 Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
9
ai-service/Dockerfile
Normal file
9
ai-service/Dockerfile
Normal file
@@ -0,0 +1,9 @@
|
||||
FROM python:3.11-slim
|
||||
WORKDIR /app
|
||||
RUN apt-get update && apt-get install -y gcc build-essential && rm -rf /var/lib/apt/lists/*
|
||||
COPY requirements.txt .
|
||||
RUN pip install --no-cache-dir -r requirements.txt
|
||||
COPY . .
|
||||
RUN mkdir -p /app/data/chroma
|
||||
EXPOSE 8000
|
||||
CMD ["uvicorn", "main:app", "--host", "0.0.0.0", "--port", "8000"]
|
||||
24
ai-service/config.py
Normal file
24
ai-service/config.py
Normal file
@@ -0,0 +1,24 @@
|
||||
import os
|
||||
|
||||
|
||||
class Settings:
|
||||
OLLAMA_BASE_URL: str = os.getenv("OLLAMA_BASE_URL", "http://100.111.160.84:11434")
|
||||
OLLAMA_TEXT_MODEL: str = os.getenv("OLLAMA_TEXT_MODEL", "qwen2.5:14b-instruct-q4_K_M")
|
||||
OLLAMA_EMBED_MODEL: str = os.getenv("OLLAMA_EMBED_MODEL", "bge-m3")
|
||||
OLLAMA_TIMEOUT: int = int(os.getenv("OLLAMA_TIMEOUT", "120"))
|
||||
|
||||
DB_HOST: str = os.getenv("DB_HOST", "mariadb")
|
||||
DB_PORT: int = int(os.getenv("DB_PORT", "3306"))
|
||||
DB_USER: str = os.getenv("DB_USER", "hyungi_user")
|
||||
DB_PASSWORD: str = os.getenv("DB_PASSWORD", "")
|
||||
DB_NAME: str = os.getenv("DB_NAME", "hyungi")
|
||||
|
||||
SECRET_KEY: str = os.getenv("SECRET_KEY", "")
|
||||
ALGORITHM: str = "HS256"
|
||||
|
||||
SYSTEM1_API_URL: str = os.getenv("SYSTEM1_API_URL", "http://system1-api:3005")
|
||||
CHROMA_PERSIST_DIR: str = os.getenv("CHROMA_PERSIST_DIR", "/app/data/chroma")
|
||||
METADATA_DB_PATH: str = os.getenv("METADATA_DB_PATH", "/app/data/metadata.db")
|
||||
|
||||
|
||||
settings = Settings()
|
||||
0
ai-service/db/__init__.py
Normal file
0
ai-service/db/__init__.py
Normal file
39
ai-service/db/metadata_store.py
Normal file
39
ai-service/db/metadata_store.py
Normal file
@@ -0,0 +1,39 @@
|
||||
import sqlite3
|
||||
from config import settings
|
||||
|
||||
|
||||
class MetadataStore:
|
||||
def __init__(self):
|
||||
self.db_path = settings.METADATA_DB_PATH
|
||||
|
||||
def initialize(self):
|
||||
conn = sqlite3.connect(self.db_path)
|
||||
conn.execute(
|
||||
"CREATE TABLE IF NOT EXISTS sync_state ("
|
||||
" key TEXT PRIMARY KEY,"
|
||||
" value TEXT"
|
||||
")"
|
||||
)
|
||||
conn.commit()
|
||||
conn.close()
|
||||
|
||||
def get_last_synced_id(self) -> int:
|
||||
conn = sqlite3.connect(self.db_path)
|
||||
cur = conn.execute(
|
||||
"SELECT value FROM sync_state WHERE key = 'last_synced_id'"
|
||||
)
|
||||
row = cur.fetchone()
|
||||
conn.close()
|
||||
return int(row[0]) if row else 0
|
||||
|
||||
def set_last_synced_id(self, issue_id: int):
|
||||
conn = sqlite3.connect(self.db_path)
|
||||
conn.execute(
|
||||
"INSERT OR REPLACE INTO sync_state (key, value) VALUES ('last_synced_id', ?)",
|
||||
(str(issue_id),),
|
||||
)
|
||||
conn.commit()
|
||||
conn.close()
|
||||
|
||||
|
||||
metadata_store = MetadataStore()
|
||||
76
ai-service/db/vector_store.py
Normal file
76
ai-service/db/vector_store.py
Normal file
@@ -0,0 +1,76 @@
|
||||
import chromadb
|
||||
from config import settings
|
||||
|
||||
|
||||
class VectorStore:
|
||||
def __init__(self):
|
||||
self.client = None
|
||||
self.collection = None
|
||||
|
||||
def initialize(self):
|
||||
self.client = chromadb.PersistentClient(path=settings.CHROMA_PERSIST_DIR)
|
||||
self.collection = self.client.get_or_create_collection(
|
||||
name="qc_issues",
|
||||
metadata={"hnsw:space": "cosine"},
|
||||
)
|
||||
|
||||
def upsert(
|
||||
self,
|
||||
doc_id: str,
|
||||
document: str,
|
||||
embedding: list[float],
|
||||
metadata: dict = None,
|
||||
):
|
||||
self.collection.upsert(
|
||||
ids=[doc_id],
|
||||
documents=[document],
|
||||
embeddings=[embedding],
|
||||
metadatas=[metadata] if metadata else None,
|
||||
)
|
||||
|
||||
def query(
|
||||
self,
|
||||
embedding: list[float],
|
||||
n_results: int = 5,
|
||||
where: dict = None,
|
||||
) -> list[dict]:
|
||||
kwargs = {
|
||||
"query_embeddings": [embedding],
|
||||
"n_results": n_results,
|
||||
"include": ["documents", "metadatas", "distances"],
|
||||
}
|
||||
if where:
|
||||
kwargs["where"] = where
|
||||
try:
|
||||
results = self.collection.query(**kwargs)
|
||||
except Exception:
|
||||
return []
|
||||
|
||||
items = []
|
||||
if results and results["ids"] and results["ids"][0]:
|
||||
for i, doc_id in enumerate(results["ids"][0]):
|
||||
item = {
|
||||
"id": doc_id,
|
||||
"document": results["documents"][0][i] if results["documents"] else "",
|
||||
"distance": results["distances"][0][i] if results["distances"] else 0,
|
||||
"metadata": results["metadatas"][0][i] if results["metadatas"] else {},
|
||||
}
|
||||
# cosine distance → similarity
|
||||
item["similarity"] = round(1 - item["distance"], 4)
|
||||
items.append(item)
|
||||
return items
|
||||
|
||||
def delete(self, doc_id: str):
|
||||
self.collection.delete(ids=[doc_id])
|
||||
|
||||
def count(self) -> int:
|
||||
return self.collection.count()
|
||||
|
||||
def stats(self) -> dict:
|
||||
return {
|
||||
"total_documents": self.count(),
|
||||
"collection_name": "qc_issues",
|
||||
}
|
||||
|
||||
|
||||
vector_store = VectorStore()
|
||||
41
ai-service/main.py
Normal file
41
ai-service/main.py
Normal file
@@ -0,0 +1,41 @@
|
||||
from contextlib import asynccontextmanager
|
||||
from fastapi import FastAPI
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
|
||||
from routers import health, embeddings, classification, daily_report, rag
|
||||
from db.vector_store import vector_store
|
||||
from db.metadata_store import metadata_store
|
||||
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
vector_store.initialize()
|
||||
metadata_store.initialize()
|
||||
yield
|
||||
|
||||
|
||||
app = FastAPI(
|
||||
title="TK AI Service",
|
||||
description="AI 서비스 (유사 검색, 분류, 보고서)",
|
||||
version="1.0.0",
|
||||
lifespan=lifespan,
|
||||
)
|
||||
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["*"],
|
||||
allow_credentials=False,
|
||||
allow_methods=["*"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
app.include_router(health.router, prefix="/api/ai")
|
||||
app.include_router(embeddings.router, prefix="/api/ai")
|
||||
app.include_router(classification.router, prefix="/api/ai")
|
||||
app.include_router(daily_report.router, prefix="/api/ai")
|
||||
app.include_router(rag.router, prefix="/api/ai")
|
||||
|
||||
|
||||
@app.get("/")
|
||||
async def root():
|
||||
return {"message": "TK AI Service", "version": "1.0.0"}
|
||||
18
ai-service/prompts/classify_issue.txt
Normal file
18
ai-service/prompts/classify_issue.txt
Normal file
@@ -0,0 +1,18 @@
|
||||
당신은 공장 품질관리(QC) 전문가입니다. 아래 부적합 신고 내용을 분석하여 판별하세요.
|
||||
|
||||
부적합 내용:
|
||||
{description}
|
||||
|
||||
상세 내용:
|
||||
{detail_notes}
|
||||
|
||||
다음 JSON 형식으로만 응답하세요:
|
||||
{{
|
||||
"category": "material_missing|design_error|incoming_defect|inspection_miss|기타",
|
||||
"category_confidence": 0.0~1.0,
|
||||
"responsible_department": "production|quality|purchasing|design|sales",
|
||||
"department_confidence": 0.0~1.0,
|
||||
"severity": "low|medium|high|critical",
|
||||
"summary": "한줄 요약 (30자 이내)",
|
||||
"reasoning": "판단 근거 (2-3문장)"
|
||||
}}
|
||||
22
ai-service/prompts/daily_report.txt
Normal file
22
ai-service/prompts/daily_report.txt
Normal file
@@ -0,0 +1,22 @@
|
||||
당신은 공장 관리 보고서 작성자입니다. 아래 데이터를 바탕으로 일일 브리핑을 작성하세요.
|
||||
|
||||
날짜: {date}
|
||||
|
||||
[근태 현황]
|
||||
{attendance_data}
|
||||
|
||||
[작업 현황]
|
||||
{work_report_data}
|
||||
|
||||
[부적합 현황]
|
||||
{qc_issue_data}
|
||||
|
||||
[순회점검 현황]
|
||||
{patrol_data}
|
||||
|
||||
다음 형식으로 작성하세요:
|
||||
|
||||
1. 오늘의 요약 (2-3문장)
|
||||
2. 주요 이슈 및 관심사항
|
||||
3. 부적합 현황 (신규/진행/지연)
|
||||
4. 내일 주의사항
|
||||
23
ai-service/prompts/rag_classify.txt
Normal file
23
ai-service/prompts/rag_classify.txt
Normal file
@@ -0,0 +1,23 @@
|
||||
당신은 공장 품질관리(QC) 전문가입니다. 아래 부적합 신고를 분류하세요.
|
||||
|
||||
[신고 내용]
|
||||
{description}
|
||||
|
||||
[상세 내용]
|
||||
{detail_notes}
|
||||
|
||||
[참고: 과거 유사 사례]
|
||||
{retrieved_cases}
|
||||
|
||||
위 과거 사례의 분류 패턴을 참고하여, 현재 부적합을 판별하세요.
|
||||
|
||||
다음 JSON 형식으로만 응답하세요:
|
||||
{{
|
||||
"category": "material_missing|design_error|incoming_defect|inspection_miss|기타",
|
||||
"category_confidence": 0.0~1.0,
|
||||
"responsible_department": "production|quality|purchasing|design|sales",
|
||||
"department_confidence": 0.0~1.0,
|
||||
"severity": "low|medium|high|critical",
|
||||
"summary": "한줄 요약 (30자 이내)",
|
||||
"reasoning": "판단 근거 — 과거 사례 참고 내용 포함 (2-3문장)"
|
||||
}}
|
||||
16
ai-service/prompts/rag_pattern.txt
Normal file
16
ai-service/prompts/rag_pattern.txt
Normal file
@@ -0,0 +1,16 @@
|
||||
당신은 공장 품질관리(QC) 데이터 분석가입니다. 아래 부적합에 대해 패턴을 분석하세요.
|
||||
|
||||
[분석 대상]
|
||||
{description}
|
||||
|
||||
[유사 부적합 {total_similar}건]
|
||||
{retrieved_cases}
|
||||
|
||||
다음을 분석하세요:
|
||||
|
||||
1. **반복 여부**: 이 문제가 과거에도 발생했는지, 반복 빈도는 어느 정도인지
|
||||
2. **공통 패턴**: 유사 사례들의 공통 원인, 공통 부서, 공통 시기 등
|
||||
3. **근본 원인 추정**: 반복되는 원인이 있다면 근본 원인은 무엇인지
|
||||
4. **개선 제안**: 재발 방지를 위한 구조적 개선 방안
|
||||
|
||||
데이터 기반으로 객관적으로 분석하세요.
|
||||
14
ai-service/prompts/rag_qa.txt
Normal file
14
ai-service/prompts/rag_qa.txt
Normal file
@@ -0,0 +1,14 @@
|
||||
당신은 공장 품질관리(QC) 데이터 분석가입니다. 아래 질문에 대해 과거 부적합 데이터를 기반으로 답변하세요.
|
||||
|
||||
[질문]
|
||||
{question}
|
||||
|
||||
[관련 부적합 데이터]
|
||||
{retrieved_cases}
|
||||
|
||||
위 데이터를 근거로 질문에 답변하세요.
|
||||
- 제공된 데이터를 적극적으로 활용하여 답변하세요
|
||||
- 관련 사례를 구체적으로 인용하며 분석하세요
|
||||
- 패턴이나 공통점이 있다면 정리하세요
|
||||
- 숫자나 통계가 있다면 포함하세요
|
||||
- 간결하되 유용한 답변을 하세요
|
||||
18
ai-service/prompts/rag_suggest_solution.txt
Normal file
18
ai-service/prompts/rag_suggest_solution.txt
Normal file
@@ -0,0 +1,18 @@
|
||||
당신은 공장 품질관리(QC) 전문가입니다. 아래 부적합 이슈에 대한 해결방안을 제안하세요.
|
||||
|
||||
[현재 부적합]
|
||||
분류: {category}
|
||||
내용: {description}
|
||||
상세: {detail_notes}
|
||||
|
||||
[과거 유사 사례]
|
||||
{retrieved_cases}
|
||||
|
||||
위 과거 사례들을 참고하여 다음을 제안하세요:
|
||||
|
||||
1. **권장 해결방안**: 과거 유사 사례에서 효과적이었던 해결 방법을 기반으로 구체적인 조치를 제안
|
||||
2. **예상 원인**: 유사 사례에서 확인된 원인 패턴을 바탕으로 가능한 원인 분석
|
||||
3. **담당 부서**: 어느 부서에서 처리해야 하는지
|
||||
4. **주의사항**: 과거 사례에서 배운 교훈이나 주의할 점
|
||||
|
||||
간결하고 실용적으로 작성하세요. 과거 사례가 없는 부분은 일반적인 QC 지식으로 보완하세요.
|
||||
17
ai-service/prompts/summarize_issue.txt
Normal file
17
ai-service/prompts/summarize_issue.txt
Normal file
@@ -0,0 +1,17 @@
|
||||
당신은 공장 품질관리(QC) 전문가입니다. 아래 부적합 이슈를 간결하게 요약하세요.
|
||||
|
||||
부적합 내용:
|
||||
{description}
|
||||
|
||||
상세 내용:
|
||||
{detail_notes}
|
||||
|
||||
해결 방법:
|
||||
{solution}
|
||||
|
||||
다음 JSON 형식으로만 응답하세요:
|
||||
{{
|
||||
"summary": "핵심 요약 (50자 이내)",
|
||||
"key_points": ["요점1", "요점2", "요점3"],
|
||||
"suggested_action": "권장 조치사항 (선택)"
|
||||
}}
|
||||
10
ai-service/requirements.txt
Normal file
10
ai-service/requirements.txt
Normal file
@@ -0,0 +1,10 @@
|
||||
fastapi==0.104.1
|
||||
uvicorn[standard]==0.24.0
|
||||
httpx==0.27.0
|
||||
chromadb==0.4.22
|
||||
numpy==1.26.2
|
||||
pydantic==2.5.0
|
||||
pydantic-settings==2.1.0
|
||||
python-jose[cryptography]==3.3.0
|
||||
pymysql==1.1.0
|
||||
sqlalchemy==2.0.23
|
||||
0
ai-service/routers/__init__.py
Normal file
0
ai-service/routers/__init__.py
Normal file
47
ai-service/routers/classification.py
Normal file
47
ai-service/routers/classification.py
Normal file
@@ -0,0 +1,47 @@
|
||||
from fastapi import APIRouter
|
||||
from pydantic import BaseModel
|
||||
from services.classification_service import (
|
||||
classify_issue,
|
||||
summarize_issue,
|
||||
classify_and_summarize,
|
||||
)
|
||||
|
||||
router = APIRouter(tags=["classification"])
|
||||
|
||||
|
||||
class ClassifyRequest(BaseModel):
|
||||
description: str
|
||||
detail_notes: str = ""
|
||||
|
||||
|
||||
class SummarizeRequest(BaseModel):
|
||||
description: str
|
||||
detail_notes: str = ""
|
||||
solution: str = ""
|
||||
|
||||
|
||||
@router.post("/classify")
|
||||
async def classify(req: ClassifyRequest):
|
||||
try:
|
||||
result = await classify_issue(req.description, req.detail_notes)
|
||||
return {"available": True, **result}
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/summarize")
|
||||
async def summarize(req: SummarizeRequest):
|
||||
try:
|
||||
result = await summarize_issue(req.description, req.detail_notes, req.solution)
|
||||
return {"available": True, **result}
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/classify-and-summarize")
|
||||
async def classify_and_summarize_endpoint(req: ClassifyRequest):
|
||||
try:
|
||||
result = await classify_and_summarize(req.description, req.detail_notes)
|
||||
return {"available": True, **result}
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
33
ai-service/routers/daily_report.py
Normal file
33
ai-service/routers/daily_report.py
Normal file
@@ -0,0 +1,33 @@
|
||||
from fastapi import APIRouter, Request
|
||||
from pydantic import BaseModel
|
||||
from services.report_service import generate_daily_report
|
||||
from datetime import date
|
||||
|
||||
router = APIRouter(tags=["daily_report"])
|
||||
|
||||
|
||||
class DailyReportRequest(BaseModel):
|
||||
date: str | None = None
|
||||
project_id: int | None = None
|
||||
|
||||
|
||||
@router.post("/report/daily")
|
||||
async def daily_report(req: DailyReportRequest, request: Request):
|
||||
report_date = req.date or date.today().isoformat()
|
||||
token = request.headers.get("authorization", "").replace("Bearer ", "")
|
||||
try:
|
||||
result = await generate_daily_report(report_date, req.project_id, token)
|
||||
return {"available": True, **result}
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/report/preview")
|
||||
async def report_preview(req: DailyReportRequest, request: Request):
|
||||
report_date = req.date or date.today().isoformat()
|
||||
token = request.headers.get("authorization", "").replace("Bearer ", "")
|
||||
try:
|
||||
result = await generate_daily_report(report_date, req.project_id, token)
|
||||
return {"available": True, "preview": True, **result}
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
77
ai-service/routers/embeddings.py
Normal file
77
ai-service/routers/embeddings.py
Normal file
@@ -0,0 +1,77 @@
|
||||
from fastapi import APIRouter, BackgroundTasks, Query
|
||||
from pydantic import BaseModel
|
||||
from services.embedding_service import (
|
||||
sync_all_issues,
|
||||
sync_single_issue,
|
||||
sync_incremental,
|
||||
search_similar_by_id,
|
||||
search_similar_by_text,
|
||||
)
|
||||
from db.vector_store import vector_store
|
||||
|
||||
router = APIRouter(tags=["embeddings"])
|
||||
|
||||
|
||||
class SyncSingleRequest(BaseModel):
|
||||
issue_id: int
|
||||
|
||||
|
||||
class SearchRequest(BaseModel):
|
||||
query: str
|
||||
n_results: int = 5
|
||||
project_id: int | None = None
|
||||
category: str | None = None
|
||||
|
||||
|
||||
@router.post("/embeddings/sync")
|
||||
async def sync_embeddings(background_tasks: BackgroundTasks):
|
||||
background_tasks.add_task(sync_all_issues)
|
||||
return {"status": "sync_started", "message": "전체 임베딩 동기화가 시작되었습니다"}
|
||||
|
||||
|
||||
@router.post("/embeddings/sync-full")
|
||||
async def sync_embeddings_full():
|
||||
result = await sync_all_issues()
|
||||
return {"status": "completed", **result}
|
||||
|
||||
|
||||
@router.post("/embeddings/sync-single")
|
||||
async def sync_single(req: SyncSingleRequest):
|
||||
result = await sync_single_issue(req.issue_id)
|
||||
return result
|
||||
|
||||
|
||||
@router.post("/embeddings/sync-incremental")
|
||||
async def sync_incr():
|
||||
result = await sync_incremental()
|
||||
return result
|
||||
|
||||
|
||||
@router.get("/similar/{issue_id}")
|
||||
async def get_similar(issue_id: int, n_results: int = Query(default=5, le=20)):
|
||||
try:
|
||||
results = await search_similar_by_id(issue_id, n_results)
|
||||
return {"available": True, "results": results, "query_issue_id": issue_id}
|
||||
except Exception as e:
|
||||
return {"available": False, "results": [], "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/similar/search")
|
||||
async def search_similar(req: SearchRequest):
|
||||
filters = {}
|
||||
if req.project_id is not None:
|
||||
filters["project_id"] = str(req.project_id)
|
||||
if req.category:
|
||||
filters["category"] = req.category
|
||||
try:
|
||||
results = await search_similar_by_text(
|
||||
req.query, req.n_results, filters or None
|
||||
)
|
||||
return {"available": True, "results": results}
|
||||
except Exception as e:
|
||||
return {"available": False, "results": [], "error": str(e)}
|
||||
|
||||
|
||||
@router.get("/embeddings/stats")
|
||||
async def embedding_stats():
|
||||
return vector_store.stats()
|
||||
21
ai-service/routers/health.py
Normal file
21
ai-service/routers/health.py
Normal file
@@ -0,0 +1,21 @@
|
||||
from fastapi import APIRouter
|
||||
from services.ollama_client import ollama_client
|
||||
from db.vector_store import vector_store
|
||||
|
||||
router = APIRouter(tags=["health"])
|
||||
|
||||
|
||||
@router.get("/health")
|
||||
async def health_check():
|
||||
ollama_status = await ollama_client.check_health()
|
||||
return {
|
||||
"status": "ok",
|
||||
"service": "tk-ai-service",
|
||||
"ollama": ollama_status,
|
||||
"embeddings": vector_store.stats(),
|
||||
}
|
||||
|
||||
|
||||
@router.get("/models")
|
||||
async def list_models():
|
||||
return await ollama_client.check_health()
|
||||
57
ai-service/routers/rag.py
Normal file
57
ai-service/routers/rag.py
Normal file
@@ -0,0 +1,57 @@
|
||||
from fastapi import APIRouter
|
||||
from pydantic import BaseModel
|
||||
from services.rag_service import (
|
||||
rag_suggest_solution,
|
||||
rag_ask,
|
||||
rag_analyze_pattern,
|
||||
rag_classify_with_context,
|
||||
)
|
||||
|
||||
router = APIRouter(tags=["rag"])
|
||||
|
||||
|
||||
class AskRequest(BaseModel):
|
||||
question: str
|
||||
project_id: int | None = None
|
||||
|
||||
|
||||
class PatternRequest(BaseModel):
|
||||
description: str
|
||||
n_results: int = 10
|
||||
|
||||
|
||||
class ClassifyRequest(BaseModel):
|
||||
description: str
|
||||
detail_notes: str = ""
|
||||
|
||||
|
||||
@router.post("/rag/suggest-solution/{issue_id}")
|
||||
async def suggest_solution(issue_id: int):
|
||||
try:
|
||||
return await rag_suggest_solution(issue_id)
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/rag/ask")
|
||||
async def ask_question(req: AskRequest):
|
||||
try:
|
||||
return await rag_ask(req.question, req.project_id)
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/rag/pattern")
|
||||
async def analyze_pattern(req: PatternRequest):
|
||||
try:
|
||||
return await rag_analyze_pattern(req.description, req.n_results)
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
|
||||
|
||||
@router.post("/rag/classify")
|
||||
async def classify_with_rag(req: ClassifyRequest):
|
||||
try:
|
||||
return await rag_classify_with_context(req.description, req.detail_notes)
|
||||
except Exception as e:
|
||||
return {"available": False, "error": str(e)}
|
||||
0
ai-service/services/__init__.py
Normal file
0
ai-service/services/__init__.py
Normal file
60
ai-service/services/classification_service.py
Normal file
60
ai-service/services/classification_service.py
Normal file
@@ -0,0 +1,60 @@
|
||||
import json
|
||||
from services.ollama_client import ollama_client
|
||||
from config import settings
|
||||
|
||||
|
||||
CLASSIFY_PROMPT_PATH = "prompts/classify_issue.txt"
|
||||
SUMMARIZE_PROMPT_PATH = "prompts/summarize_issue.txt"
|
||||
|
||||
|
||||
def _load_prompt(path: str) -> str:
|
||||
with open(path, "r", encoding="utf-8") as f:
|
||||
return f.read()
|
||||
|
||||
|
||||
async def classify_issue(description: str, detail_notes: str = "") -> dict:
|
||||
template = _load_prompt(CLASSIFY_PROMPT_PATH)
|
||||
prompt = template.format(
|
||||
description=description or "",
|
||||
detail_notes=detail_notes or "",
|
||||
)
|
||||
raw = await ollama_client.generate_text(prompt)
|
||||
try:
|
||||
start = raw.find("{")
|
||||
end = raw.rfind("}") + 1
|
||||
if start >= 0 and end > start:
|
||||
return json.loads(raw[start:end])
|
||||
except json.JSONDecodeError:
|
||||
pass
|
||||
return {"raw_response": raw, "parse_error": True}
|
||||
|
||||
|
||||
async def summarize_issue(
|
||||
description: str, detail_notes: str = "", solution: str = ""
|
||||
) -> dict:
|
||||
template = _load_prompt(SUMMARIZE_PROMPT_PATH)
|
||||
prompt = template.format(
|
||||
description=description or "",
|
||||
detail_notes=detail_notes or "",
|
||||
solution=solution or "",
|
||||
)
|
||||
raw = await ollama_client.generate_text(prompt)
|
||||
try:
|
||||
start = raw.find("{")
|
||||
end = raw.rfind("}") + 1
|
||||
if start >= 0 and end > start:
|
||||
return json.loads(raw[start:end])
|
||||
except json.JSONDecodeError:
|
||||
pass
|
||||
return {"summary": raw.strip()}
|
||||
|
||||
|
||||
async def classify_and_summarize(
|
||||
description: str, detail_notes: str = ""
|
||||
) -> dict:
|
||||
classification = await classify_issue(description, detail_notes)
|
||||
summary_result = await summarize_issue(description, detail_notes)
|
||||
return {
|
||||
"classification": classification,
|
||||
"summary": summary_result.get("summary", ""),
|
||||
}
|
||||
97
ai-service/services/db_client.py
Normal file
97
ai-service/services/db_client.py
Normal file
@@ -0,0 +1,97 @@
|
||||
from urllib.parse import quote_plus
|
||||
|
||||
from sqlalchemy import create_engine, text
|
||||
from config import settings
|
||||
|
||||
|
||||
def get_engine():
|
||||
password = quote_plus(settings.DB_PASSWORD)
|
||||
url = (
|
||||
f"mysql+pymysql://{settings.DB_USER}:{password}"
|
||||
f"@{settings.DB_HOST}:{settings.DB_PORT}/{settings.DB_NAME}"
|
||||
)
|
||||
return create_engine(url, pool_pre_ping=True, pool_size=5)
|
||||
|
||||
|
||||
engine = get_engine()
|
||||
|
||||
|
||||
def get_all_issues() -> list[dict]:
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(
|
||||
text(
|
||||
"SELECT id, category, description, detail_notes, "
|
||||
"final_description, final_category, solution, "
|
||||
"management_comment, cause_detail, project_id, "
|
||||
"review_status, report_date, responsible_department, "
|
||||
"location_info "
|
||||
"FROM qc_issues ORDER BY id"
|
||||
)
|
||||
)
|
||||
return [dict(row._mapping) for row in result]
|
||||
|
||||
|
||||
def get_issue_by_id(issue_id: int) -> dict | None:
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(
|
||||
text(
|
||||
"SELECT id, category, description, detail_notes, "
|
||||
"final_description, final_category, solution, "
|
||||
"management_comment, cause_detail, project_id, "
|
||||
"review_status, report_date, responsible_department, "
|
||||
"location_info "
|
||||
"FROM qc_issues WHERE id = :id"
|
||||
),
|
||||
{"id": issue_id},
|
||||
)
|
||||
row = result.fetchone()
|
||||
return dict(row._mapping) if row else None
|
||||
|
||||
|
||||
def get_issues_since(last_id: int) -> list[dict]:
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(
|
||||
text(
|
||||
"SELECT id, category, description, detail_notes, "
|
||||
"final_description, final_category, solution, "
|
||||
"management_comment, cause_detail, project_id, "
|
||||
"review_status, report_date, responsible_department, "
|
||||
"location_info "
|
||||
"FROM qc_issues WHERE id > :last_id ORDER BY id"
|
||||
),
|
||||
{"last_id": last_id},
|
||||
)
|
||||
return [dict(row._mapping) for row in result]
|
||||
|
||||
|
||||
def get_daily_qc_stats(date_str: str) -> dict:
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(
|
||||
text(
|
||||
"SELECT "
|
||||
" COUNT(*) as total, "
|
||||
" SUM(CASE WHEN DATE(report_date) = :d THEN 1 ELSE 0 END) as new_today, "
|
||||
" SUM(CASE WHEN review_status = 'in_progress' THEN 1 ELSE 0 END) as in_progress, "
|
||||
" SUM(CASE WHEN review_status = 'completed' THEN 1 ELSE 0 END) as completed, "
|
||||
" SUM(CASE WHEN review_status = 'pending_review' THEN 1 ELSE 0 END) as pending "
|
||||
"FROM qc_issues"
|
||||
),
|
||||
{"d": date_str},
|
||||
)
|
||||
row = result.fetchone()
|
||||
return dict(row._mapping) if row else {}
|
||||
|
||||
|
||||
def get_issues_for_date(date_str: str) -> list[dict]:
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(
|
||||
text(
|
||||
"SELECT id, category, description, detail_notes, "
|
||||
"review_status, responsible_department, solution "
|
||||
"FROM qc_issues "
|
||||
"WHERE DATE(report_date) = :d "
|
||||
"ORDER BY id"
|
||||
),
|
||||
{"d": date_str},
|
||||
)
|
||||
return [dict(row._mapping) for row in result]
|
||||
144
ai-service/services/embedding_service.py
Normal file
144
ai-service/services/embedding_service.py
Normal file
@@ -0,0 +1,144 @@
|
||||
from services.ollama_client import ollama_client
|
||||
from db.vector_store import vector_store
|
||||
from db.metadata_store import metadata_store
|
||||
from services.db_client import get_all_issues, get_issue_by_id, get_issues_since
|
||||
|
||||
|
||||
def build_document_text(issue: dict) -> str:
|
||||
parts = []
|
||||
if issue.get("description"):
|
||||
parts.append(issue["description"])
|
||||
if issue.get("final_description"):
|
||||
parts.append(issue["final_description"])
|
||||
if issue.get("detail_notes"):
|
||||
parts.append(issue["detail_notes"])
|
||||
if issue.get("solution"):
|
||||
parts.append(f"해결: {issue['solution']}")
|
||||
if issue.get("management_comment"):
|
||||
parts.append(f"의견: {issue['management_comment']}")
|
||||
if issue.get("cause_detail"):
|
||||
parts.append(f"원인: {issue['cause_detail']}")
|
||||
return " ".join(parts)
|
||||
|
||||
|
||||
def build_metadata(issue: dict) -> dict:
|
||||
meta = {"issue_id": issue["id"]}
|
||||
for key in [
|
||||
"category", "project_id", "review_status",
|
||||
"responsible_department", "location_info",
|
||||
]:
|
||||
val = issue.get(key)
|
||||
if val is not None:
|
||||
meta[key] = str(val)
|
||||
rd = issue.get("report_date")
|
||||
if rd:
|
||||
meta["report_date"] = str(rd)[:10]
|
||||
meta["has_solution"] = "true" if issue.get("solution") else "false"
|
||||
return meta
|
||||
|
||||
|
||||
async def sync_all_issues() -> dict:
|
||||
issues = get_all_issues()
|
||||
synced = 0
|
||||
skipped = 0
|
||||
for issue in issues:
|
||||
doc_text = build_document_text(issue)
|
||||
if not doc_text.strip():
|
||||
skipped += 1
|
||||
continue
|
||||
try:
|
||||
embedding = await ollama_client.generate_embedding(doc_text)
|
||||
vector_store.upsert(
|
||||
doc_id=f"issue_{issue['id']}",
|
||||
document=doc_text,
|
||||
embedding=embedding,
|
||||
metadata=build_metadata(issue),
|
||||
)
|
||||
synced += 1
|
||||
except Exception as e:
|
||||
skipped += 1
|
||||
if issues:
|
||||
max_id = max(i["id"] for i in issues)
|
||||
metadata_store.set_last_synced_id(max_id)
|
||||
return {"synced": synced, "skipped": skipped, "total": len(issues)}
|
||||
|
||||
|
||||
async def sync_single_issue(issue_id: int) -> dict:
|
||||
issue = get_issue_by_id(issue_id)
|
||||
if not issue:
|
||||
return {"status": "not_found"}
|
||||
doc_text = build_document_text(issue)
|
||||
if not doc_text.strip():
|
||||
return {"status": "empty_text"}
|
||||
embedding = await ollama_client.generate_embedding(doc_text)
|
||||
vector_store.upsert(
|
||||
doc_id=f"issue_{issue['id']}",
|
||||
document=doc_text,
|
||||
embedding=embedding,
|
||||
metadata=build_metadata(issue),
|
||||
)
|
||||
return {"status": "synced", "issue_id": issue_id}
|
||||
|
||||
|
||||
async def sync_incremental() -> dict:
|
||||
last_id = metadata_store.get_last_synced_id()
|
||||
issues = get_issues_since(last_id)
|
||||
synced = 0
|
||||
for issue in issues:
|
||||
doc_text = build_document_text(issue)
|
||||
if not doc_text.strip():
|
||||
continue
|
||||
try:
|
||||
embedding = await ollama_client.generate_embedding(doc_text)
|
||||
vector_store.upsert(
|
||||
doc_id=f"issue_{issue['id']}",
|
||||
document=doc_text,
|
||||
embedding=embedding,
|
||||
metadata=build_metadata(issue),
|
||||
)
|
||||
synced += 1
|
||||
except Exception:
|
||||
pass
|
||||
if issues:
|
||||
max_id = max(i["id"] for i in issues)
|
||||
metadata_store.set_last_synced_id(max_id)
|
||||
return {"synced": synced, "new_issues": len(issues)}
|
||||
|
||||
|
||||
async def search_similar_by_id(issue_id: int, n_results: int = 5) -> list[dict]:
|
||||
issue = get_issue_by_id(issue_id)
|
||||
if not issue:
|
||||
return []
|
||||
doc_text = build_document_text(issue)
|
||||
if not doc_text.strip():
|
||||
return []
|
||||
embedding = await ollama_client.generate_embedding(doc_text)
|
||||
results = vector_store.query(
|
||||
embedding=embedding,
|
||||
n_results=n_results + 1,
|
||||
)
|
||||
# exclude self
|
||||
filtered = []
|
||||
for r in results:
|
||||
if r["id"] != f"issue_{issue_id}":
|
||||
filtered.append(r)
|
||||
return filtered[:n_results]
|
||||
|
||||
|
||||
async def search_similar_by_text(query: str, n_results: int = 5, filters: dict = None) -> list[dict]:
|
||||
embedding = await ollama_client.generate_embedding(query)
|
||||
where = None
|
||||
if filters:
|
||||
conditions = []
|
||||
for k, v in filters.items():
|
||||
if v is not None:
|
||||
conditions.append({k: str(v)})
|
||||
if len(conditions) == 1:
|
||||
where = conditions[0]
|
||||
elif len(conditions) > 1:
|
||||
where = {"$and": conditions}
|
||||
return vector_store.query(
|
||||
embedding=embedding,
|
||||
n_results=n_results,
|
||||
where=where,
|
||||
)
|
||||
57
ai-service/services/ollama_client.py
Normal file
57
ai-service/services/ollama_client.py
Normal file
@@ -0,0 +1,57 @@
|
||||
import httpx
|
||||
from config import settings
|
||||
|
||||
|
||||
class OllamaClient:
|
||||
def __init__(self):
|
||||
self.base_url = settings.OLLAMA_BASE_URL
|
||||
self.timeout = httpx.Timeout(float(settings.OLLAMA_TIMEOUT), connect=10.0)
|
||||
|
||||
async def generate_embedding(self, text: str) -> list[float]:
|
||||
async with httpx.AsyncClient(timeout=self.timeout) as client:
|
||||
response = await client.post(
|
||||
f"{self.base_url}/api/embeddings",
|
||||
json={"model": settings.OLLAMA_EMBED_MODEL, "prompt": text},
|
||||
)
|
||||
response.raise_for_status()
|
||||
return response.json()["embedding"]
|
||||
|
||||
async def batch_embeddings(self, texts: list[str]) -> list[list[float]]:
|
||||
results = []
|
||||
for text in texts:
|
||||
emb = await self.generate_embedding(text)
|
||||
results.append(emb)
|
||||
return results
|
||||
|
||||
async def generate_text(self, prompt: str, system: str = None) -> str:
|
||||
messages = []
|
||||
if system:
|
||||
messages.append({"role": "system", "content": system})
|
||||
messages.append({"role": "user", "content": prompt})
|
||||
async with httpx.AsyncClient(timeout=self.timeout) as client:
|
||||
response = await client.post(
|
||||
f"{self.base_url}/api/chat",
|
||||
json={
|
||||
"model": settings.OLLAMA_TEXT_MODEL,
|
||||
"messages": messages,
|
||||
"stream": False,
|
||||
"options": {"temperature": 0.3, "num_predict": 2048},
|
||||
},
|
||||
)
|
||||
response.raise_for_status()
|
||||
return response.json()["message"]["content"]
|
||||
|
||||
async def check_health(self) -> dict:
|
||||
try:
|
||||
async with httpx.AsyncClient(timeout=httpx.Timeout(5.0)) as client:
|
||||
response = await client.get(f"{self.base_url}/api/tags")
|
||||
models = response.json().get("models", [])
|
||||
return {
|
||||
"status": "connected",
|
||||
"models": [m["name"] for m in models],
|
||||
}
|
||||
except Exception:
|
||||
return {"status": "disconnected"}
|
||||
|
||||
|
||||
ollama_client = OllamaClient()
|
||||
164
ai-service/services/rag_service.py
Normal file
164
ai-service/services/rag_service.py
Normal file
@@ -0,0 +1,164 @@
|
||||
from services.ollama_client import ollama_client
|
||||
from services.embedding_service import search_similar_by_text, build_document_text
|
||||
from services.db_client import get_issue_by_id
|
||||
|
||||
|
||||
def _load_prompt(path: str) -> str:
|
||||
with open(path, "r", encoding="utf-8") as f:
|
||||
return f.read()
|
||||
|
||||
|
||||
def _format_retrieved_issues(results: list[dict]) -> str:
|
||||
if not results:
|
||||
return "관련 과거 사례가 없습니다."
|
||||
lines = []
|
||||
for i, r in enumerate(results, 1):
|
||||
meta = r.get("metadata", {})
|
||||
similarity = round(r.get("similarity", 0) * 100)
|
||||
doc = (r.get("document", ""))[:500]
|
||||
cat = meta.get("category", "")
|
||||
dept = meta.get("responsible_department", "")
|
||||
status = meta.get("review_status", "")
|
||||
has_sol = meta.get("has_solution", "false")
|
||||
date = meta.get("report_date", "")
|
||||
issue_id = meta.get("issue_id", r["id"])
|
||||
lines.append(
|
||||
f"[사례 {i}] No.{issue_id} (유사도 {similarity}%)\n"
|
||||
f" 분류: {cat} | 부서: {dept} | 상태: {status} | 날짜: {date} | 해결여부: {'O' if has_sol == 'true' else 'X'}\n"
|
||||
f" 내용: {doc}"
|
||||
)
|
||||
return "\n\n".join(lines)
|
||||
|
||||
|
||||
async def rag_suggest_solution(issue_id: int) -> dict:
|
||||
"""과거 유사 이슈의 해결 사례를 참고하여 해결방안을 제안"""
|
||||
issue = get_issue_by_id(issue_id)
|
||||
if not issue:
|
||||
return {"available": False, "error": "이슈를 찾을 수 없습니다"}
|
||||
|
||||
doc_text = build_document_text(issue)
|
||||
if not doc_text.strip():
|
||||
return {"available": False, "error": "이슈 내용이 비어있습니다"}
|
||||
|
||||
# 해결 완료된 유사 이슈 검색
|
||||
similar = await search_similar_by_text(
|
||||
doc_text, n_results=5, filters={"has_solution": "true"}
|
||||
)
|
||||
# 해결 안 된 것도 포함 (참고용)
|
||||
if len(similar) < 3:
|
||||
all_similar = await search_similar_by_text(doc_text, n_results=5)
|
||||
seen = {r["id"] for r in similar}
|
||||
for r in all_similar:
|
||||
if r["id"] not in seen:
|
||||
similar.append(r)
|
||||
if len(similar) >= 5:
|
||||
break
|
||||
|
||||
context = _format_retrieved_issues(similar)
|
||||
template = _load_prompt("prompts/rag_suggest_solution.txt")
|
||||
prompt = template.format(
|
||||
description=issue.get("description", ""),
|
||||
detail_notes=issue.get("detail_notes", ""),
|
||||
category=issue.get("category", ""),
|
||||
retrieved_cases=context,
|
||||
)
|
||||
|
||||
response = await ollama_client.generate_text(prompt)
|
||||
return {
|
||||
"available": True,
|
||||
"issue_id": issue_id,
|
||||
"suggestion": response,
|
||||
"referenced_issues": [
|
||||
{
|
||||
"id": r.get("metadata", {}).get("issue_id", r["id"]),
|
||||
"similarity": round(r.get("similarity", 0) * 100),
|
||||
"has_solution": r.get("metadata", {}).get("has_solution", "false") == "true",
|
||||
}
|
||||
for r in similar
|
||||
],
|
||||
}
|
||||
|
||||
|
||||
async def rag_ask(question: str, project_id: int = None) -> dict:
|
||||
"""부적합 데이터를 기반으로 자연어 질문에 답변"""
|
||||
# 프로젝트 필터 없이 전체 데이터에서 검색 (과거 미지정 데이터 포함)
|
||||
results = await search_similar_by_text(
|
||||
question, n_results=15, filters=None
|
||||
)
|
||||
context = _format_retrieved_issues(results)
|
||||
|
||||
template = _load_prompt("prompts/rag_qa.txt")
|
||||
prompt = template.format(
|
||||
question=question,
|
||||
retrieved_cases=context,
|
||||
)
|
||||
|
||||
response = await ollama_client.generate_text(prompt)
|
||||
return {
|
||||
"available": True,
|
||||
"answer": response,
|
||||
"sources": [
|
||||
{
|
||||
"id": r.get("metadata", {}).get("issue_id", r["id"]),
|
||||
"similarity": round(r.get("similarity", 0) * 100),
|
||||
"snippet": (r.get("document", ""))[:100],
|
||||
}
|
||||
for r in results
|
||||
],
|
||||
}
|
||||
|
||||
|
||||
async def rag_analyze_pattern(description: str, n_results: int = 10) -> dict:
|
||||
"""유사 부적합 패턴 분석 — 반복되는 문제인지, 근본 원인은 무엇인지"""
|
||||
results = await search_similar_by_text(description, n_results=n_results)
|
||||
context = _format_retrieved_issues(results)
|
||||
|
||||
template = _load_prompt("prompts/rag_pattern.txt")
|
||||
prompt = template.format(
|
||||
description=description,
|
||||
retrieved_cases=context,
|
||||
total_similar=len(results),
|
||||
)
|
||||
|
||||
response = await ollama_client.generate_text(prompt)
|
||||
return {
|
||||
"available": True,
|
||||
"analysis": response,
|
||||
"similar_count": len(results),
|
||||
"sources": [
|
||||
{
|
||||
"id": r.get("metadata", {}).get("issue_id", r["id"]),
|
||||
"similarity": round(r.get("similarity", 0) * 100),
|
||||
"category": r.get("metadata", {}).get("category", ""),
|
||||
}
|
||||
for r in results
|
||||
],
|
||||
}
|
||||
|
||||
|
||||
async def rag_classify_with_context(description: str, detail_notes: str = "") -> dict:
|
||||
"""과거 사례를 참고하여 더 정확한 분류 수행 (기존 classify 강화)"""
|
||||
query = f"{description} {detail_notes}".strip()
|
||||
similar = await search_similar_by_text(query, n_results=5)
|
||||
context = _format_retrieved_issues(similar)
|
||||
|
||||
template = _load_prompt("prompts/rag_classify.txt")
|
||||
prompt = template.format(
|
||||
description=description,
|
||||
detail_notes=detail_notes,
|
||||
retrieved_cases=context,
|
||||
)
|
||||
|
||||
raw = await ollama_client.generate_text(prompt)
|
||||
import json
|
||||
try:
|
||||
start = raw.find("{")
|
||||
end = raw.rfind("}") + 1
|
||||
if start >= 0 and end > start:
|
||||
result = json.loads(raw[start:end])
|
||||
result["rag_enhanced"] = True
|
||||
result["referenced_count"] = len(similar)
|
||||
return {"available": True, **result}
|
||||
except json.JSONDecodeError:
|
||||
pass
|
||||
return {"available": True, "raw_response": raw, "rag_enhanced": True}
|
||||
122
ai-service/services/report_service.py
Normal file
122
ai-service/services/report_service.py
Normal file
@@ -0,0 +1,122 @@
|
||||
import httpx
|
||||
from services.ollama_client import ollama_client
|
||||
from services.db_client import get_daily_qc_stats, get_issues_for_date
|
||||
from config import settings
|
||||
|
||||
|
||||
REPORT_PROMPT_PATH = "prompts/daily_report.txt"
|
||||
|
||||
|
||||
def _load_prompt(path: str) -> str:
|
||||
with open(path, "r", encoding="utf-8") as f:
|
||||
return f.read()
|
||||
|
||||
|
||||
async def _fetch_system1_data(date_str: str, token: str) -> dict:
|
||||
headers = {"Authorization": f"Bearer {token}"}
|
||||
data = {"attendance": None, "work_reports": None, "patrol": None}
|
||||
try:
|
||||
async with httpx.AsyncClient(timeout=15.0) as client:
|
||||
# 근태
|
||||
try:
|
||||
r = await client.get(
|
||||
f"{settings.SYSTEM1_API_URL}/api/attendance/daily-status",
|
||||
params={"date": date_str},
|
||||
headers=headers,
|
||||
)
|
||||
if r.status_code == 200:
|
||||
data["attendance"] = r.json()
|
||||
except Exception:
|
||||
pass
|
||||
# 작업보고
|
||||
try:
|
||||
r = await client.get(
|
||||
f"{settings.SYSTEM1_API_URL}/api/daily-work-reports/summary",
|
||||
params={"date": date_str},
|
||||
headers=headers,
|
||||
)
|
||||
if r.status_code == 200:
|
||||
data["work_reports"] = r.json()
|
||||
except Exception:
|
||||
pass
|
||||
# 순회점검
|
||||
try:
|
||||
r = await client.get(
|
||||
f"{settings.SYSTEM1_API_URL}/api/patrol/today-status",
|
||||
params={"date": date_str},
|
||||
headers=headers,
|
||||
)
|
||||
if r.status_code == 200:
|
||||
data["patrol"] = r.json()
|
||||
except Exception:
|
||||
pass
|
||||
except Exception:
|
||||
pass
|
||||
return data
|
||||
|
||||
|
||||
def _format_attendance(data) -> str:
|
||||
if not data:
|
||||
return "데이터 없음"
|
||||
if isinstance(data, dict):
|
||||
parts = []
|
||||
for k, v in data.items():
|
||||
parts.append(f" {k}: {v}")
|
||||
return "\n".join(parts)
|
||||
return str(data)
|
||||
|
||||
|
||||
def _format_work_reports(data) -> str:
|
||||
if not data:
|
||||
return "데이터 없음"
|
||||
return str(data)
|
||||
|
||||
|
||||
def _format_qc_issues(issues: list[dict], stats: dict) -> str:
|
||||
lines = []
|
||||
lines.append(f"전체: {stats.get('total', 0)}건")
|
||||
lines.append(f"금일 신규: {stats.get('new_today', 0)}건")
|
||||
lines.append(f"진행중: {stats.get('in_progress', 0)}건")
|
||||
lines.append(f"완료: {stats.get('completed', 0)}건")
|
||||
lines.append(f"미검토: {stats.get('pending', 0)}건")
|
||||
if issues:
|
||||
lines.append("\n금일 신규 이슈:")
|
||||
for iss in issues[:10]:
|
||||
cat = iss.get("category", "")
|
||||
desc = (iss.get("description") or "")[:50]
|
||||
status = iss.get("review_status", "")
|
||||
lines.append(f" - [{cat}] {desc} (상태: {status})")
|
||||
return "\n".join(lines)
|
||||
|
||||
|
||||
def _format_patrol(data) -> str:
|
||||
if not data:
|
||||
return "데이터 없음"
|
||||
return str(data)
|
||||
|
||||
|
||||
async def generate_daily_report(
|
||||
date_str: str, project_id: int = None, token: str = ""
|
||||
) -> dict:
|
||||
system1_data = await _fetch_system1_data(date_str, token)
|
||||
qc_stats = get_daily_qc_stats(date_str)
|
||||
qc_issues = get_issues_for_date(date_str)
|
||||
|
||||
template = _load_prompt(REPORT_PROMPT_PATH)
|
||||
prompt = template.format(
|
||||
date=date_str,
|
||||
attendance_data=_format_attendance(system1_data["attendance"]),
|
||||
work_report_data=_format_work_reports(system1_data["work_reports"]),
|
||||
qc_issue_data=_format_qc_issues(qc_issues, qc_stats),
|
||||
patrol_data=_format_patrol(system1_data["patrol"]),
|
||||
)
|
||||
|
||||
report_text = await ollama_client.generate_text(prompt)
|
||||
return {
|
||||
"date": date_str,
|
||||
"report": report_text,
|
||||
"stats": {
|
||||
"qc": qc_stats,
|
||||
"new_issues_count": len(qc_issues),
|
||||
},
|
||||
}
|
||||
Reference in New Issue
Block a user