improve colors
This commit is contained in:
@@ -7,24 +7,19 @@ from fastapi.middleware.cors import CORSMiddleware
|
||||
from fastapi.responses import JSONResponse
|
||||
|
||||
from .core.config import settings
|
||||
from .routers.agent import router as agent_router
|
||||
from .routers.chunking import router as chunking_router
|
||||
from .routers.chunking_landingai import router as chunking_landingai_router
|
||||
from .routers.dataroom import router as dataroom_router
|
||||
|
||||
# Import routers
|
||||
from .routers.files import router as files_router
|
||||
from .routers.schemas import router as schemas_router
|
||||
from .routers.vectors import router as vectors_router
|
||||
|
||||
# from routers.ai import router as ai_router # futuro con Azure OpenAI
|
||||
|
||||
# Import config
|
||||
|
||||
|
||||
# Configurar logging
|
||||
logging.basicConfig(
|
||||
level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s"
|
||||
level=logging.WARNING, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s"
|
||||
)
|
||||
logging.getLogger("app").setLevel(logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
@@ -130,6 +125,8 @@ app.include_router(chunking_landingai_router)
|
||||
|
||||
app.include_router(dataroom_router, prefix="/api/v1")
|
||||
|
||||
app.include_router(agent_router)
|
||||
|
||||
# Router para IA
|
||||
# app.include_router(
|
||||
# ai_router,
|
||||
|
||||
24
backend/app/routers/agent.py
Normal file
24
backend/app/routers/agent.py
Normal file
@@ -0,0 +1,24 @@
|
||||
from fastapi import APIRouter
|
||||
from pydantic_ai import Agent
|
||||
from pydantic_ai.models.openai import OpenAIChatModel
|
||||
from pydantic_ai.providers.azure import AzureProvider
|
||||
from pydantic_ai.ui.vercel_ai import VercelAIAdapter
|
||||
from starlette.requests import Request
|
||||
from starlette.responses import Response
|
||||
|
||||
from app.core.config import settings
|
||||
|
||||
provider = AzureProvider(
|
||||
azure_endpoint=settings.AZURE_OPENAI_ENDPOINT,
|
||||
api_version=settings.AZURE_OPENAI_API_VERSION,
|
||||
api_key=settings.AZURE_OPENAI_API_KEY,
|
||||
)
|
||||
model = OpenAIChatModel(model_name="gpt-4o", provider=provider)
|
||||
agent = Agent(model=model)
|
||||
|
||||
router = APIRouter(prefix="/api/v1/agent", tags=["Agent"])
|
||||
|
||||
|
||||
@router.post("/chat")
|
||||
async def chat(request: Request) -> Response:
|
||||
return await VercelAIAdapter.dispatch_request(request, agent=agent)
|
||||
@@ -1,9 +1,17 @@
|
||||
from azure.storage.blob import BlobServiceClient, BlobClient, ContainerClient, generate_blob_sas, BlobSasPermissions
|
||||
from azure.core.exceptions import ResourceNotFoundError, ResourceExistsError
|
||||
from typing import List, Optional, BinaryIO
|
||||
import logging
|
||||
from datetime import datetime, timezone, timedelta
|
||||
import os
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from typing import BinaryIO, List, Optional
|
||||
|
||||
from azure.core.exceptions import ResourceExistsError, ResourceNotFoundError
|
||||
from azure.storage.blob import (
|
||||
BlobClient,
|
||||
BlobSasPermissions,
|
||||
BlobServiceClient,
|
||||
ContainerClient,
|
||||
generate_blob_sas,
|
||||
)
|
||||
|
||||
from ..core.config import settings
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
@@ -13,7 +21,7 @@ class AzureBlobService:
|
||||
"""
|
||||
Servicio para interactuar con Azure Blob Storage
|
||||
"""
|
||||
|
||||
|
||||
def __init__(self):
|
||||
"""Inicializar el cliente de Azure Blob Storage"""
|
||||
try:
|
||||
@@ -21,7 +29,9 @@ class AzureBlobService:
|
||||
settings.AZURE_STORAGE_CONNECTION_STRING
|
||||
)
|
||||
self.container_name = settings.AZURE_CONTAINER_NAME
|
||||
logger.info(f"Cliente de Azure Blob Storage inicializado para container: {self.container_name}")
|
||||
logger.info(
|
||||
f"Cliente de Azure Blob Storage inicializado para container: {self.container_name}"
|
||||
)
|
||||
|
||||
# Configurar CORS automáticamente al inicializar
|
||||
self._configure_cors()
|
||||
@@ -45,7 +55,7 @@ class AzureBlobService:
|
||||
allowed_methods=["GET", "HEAD", "OPTIONS"],
|
||||
allowed_headers=["*"],
|
||||
exposed_headers=["*"],
|
||||
max_age_in_seconds=3600
|
||||
max_age_in_seconds=3600,
|
||||
)
|
||||
|
||||
# Aplicar la configuración CORS
|
||||
@@ -55,15 +65,19 @@ class AzureBlobService:
|
||||
except Exception as e:
|
||||
# No fallar si CORS no se puede configurar (puede que ya esté configurado)
|
||||
logger.warning(f"No se pudo configurar CORS automáticamente: {e}")
|
||||
logger.warning("Asegúrate de configurar CORS manualmente en Azure Portal si es necesario")
|
||||
|
||||
logger.warning(
|
||||
"Asegúrate de configurar CORS manualmente en Azure Portal si es necesario"
|
||||
)
|
||||
|
||||
async def create_container_if_not_exists(self) -> bool:
|
||||
"""
|
||||
Crear el container si no existe
|
||||
Returns: True si se creó, False si ya existía
|
||||
"""
|
||||
try:
|
||||
container_client = self.blob_service_client.get_container_client(self.container_name)
|
||||
container_client = self.blob_service_client.get_container_client(
|
||||
self.container_name
|
||||
)
|
||||
container_client.create_container()
|
||||
logger.info(f"Container '{self.container_name}' creado exitosamente")
|
||||
return True
|
||||
@@ -73,217 +87,249 @@ class AzureBlobService:
|
||||
except Exception as e:
|
||||
logger.error(f"Error creando container: {e}")
|
||||
raise e
|
||||
|
||||
async def upload_file(self, file_data: BinaryIO, blob_name: str, tema: str = "") -> dict:
|
||||
|
||||
async def upload_file(
|
||||
self, file_data: BinaryIO, blob_name: str, tema: str = ""
|
||||
) -> dict:
|
||||
"""
|
||||
Subir un archivo a Azure Blob Storage
|
||||
|
||||
|
||||
Args:
|
||||
file_data: Datos del archivo
|
||||
blob_name: Nombre del archivo en el blob
|
||||
tema: Tema/carpeta donde guardar el archivo
|
||||
|
||||
tema: Tema/carpeta donde guardar el archivo (se normaliza a lowercase)
|
||||
|
||||
Returns:
|
||||
dict: Información del archivo subido
|
||||
"""
|
||||
try:
|
||||
# Construir la ruta completa con tema si se proporciona
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
|
||||
# Normalizar tema a lowercase para consistencia
|
||||
tema_normalized = tema.lower() if tema else ""
|
||||
|
||||
# Construir la ruta completa con tema normalizado
|
||||
full_blob_name = (
|
||||
f"{tema_normalized}/{blob_name}" if tema_normalized else blob_name
|
||||
)
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
|
||||
# Subir el archivo
|
||||
blob_client.upload_blob(file_data, overwrite=True)
|
||||
|
||||
|
||||
# Obtener propiedades del blob
|
||||
blob_properties = blob_client.get_blob_properties()
|
||||
|
||||
|
||||
logger.info(f"Archivo '{full_blob_name}' subido exitosamente")
|
||||
|
||||
|
||||
return {
|
||||
"name": blob_name,
|
||||
"full_path": full_blob_name,
|
||||
"tema": tema,
|
||||
"tema": tema_normalized,
|
||||
"size": blob_properties.size,
|
||||
"last_modified": blob_properties.last_modified,
|
||||
"url": blob_client.url
|
||||
"url": blob_client.url,
|
||||
}
|
||||
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error subiendo archivo '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
|
||||
async def download_file(self, blob_name: str, tema: str = "") -> bytes:
|
||||
"""
|
||||
Descargar un archivo de Azure Blob Storage
|
||||
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo
|
||||
tema: Tema/carpeta donde está el archivo
|
||||
|
||||
tema: Tema/carpeta donde está el archivo (búsqueda case-insensitive)
|
||||
|
||||
Returns:
|
||||
bytes: Contenido del archivo
|
||||
"""
|
||||
try:
|
||||
# Construir la ruta completa
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
|
||||
# Si se proporciona tema, buscar el archivo de manera case-insensitive
|
||||
if tema:
|
||||
full_blob_name = await self._find_blob_case_insensitive(blob_name, tema)
|
||||
else:
|
||||
full_blob_name = blob_name
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
|
||||
# Descargar el archivo
|
||||
blob_data = blob_client.download_blob()
|
||||
content = blob_data.readall()
|
||||
|
||||
|
||||
logger.info(f"Archivo '{full_blob_name}' descargado exitosamente")
|
||||
return content
|
||||
|
||||
|
||||
except ResourceNotFoundError:
|
||||
logger.error(f"Archivo '{full_blob_name}' no encontrado")
|
||||
raise FileNotFoundError(f"El archivo '{blob_name}' no existe")
|
||||
except Exception as e:
|
||||
logger.error(f"Error descargando archivo '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
|
||||
async def delete_file(self, blob_name: str, tema: str = "") -> bool:
|
||||
"""
|
||||
Eliminar un archivo de Azure Blob Storage
|
||||
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo
|
||||
tema: Tema/carpeta donde está el archivo
|
||||
|
||||
tema: Tema/carpeta donde está el archivo (búsqueda case-insensitive)
|
||||
|
||||
Returns:
|
||||
bool: True si se eliminó exitosamente
|
||||
"""
|
||||
try:
|
||||
# Construir la ruta completa
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
|
||||
# Si se proporciona tema, buscar el archivo de manera case-insensitive
|
||||
if tema:
|
||||
full_blob_name = await self._find_blob_case_insensitive(blob_name, tema)
|
||||
else:
|
||||
full_blob_name = blob_name
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
|
||||
# Eliminar el archivo
|
||||
blob_client.delete_blob()
|
||||
|
||||
|
||||
logger.info(f"Archivo '{full_blob_name}' eliminado exitosamente")
|
||||
return True
|
||||
|
||||
|
||||
except ResourceNotFoundError:
|
||||
logger.error(f"Archivo '{full_blob_name}' no encontrado para eliminar")
|
||||
raise FileNotFoundError(f"El archivo '{blob_name}' no existe")
|
||||
except Exception as e:
|
||||
logger.error(f"Error eliminando archivo '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
|
||||
async def list_files(self, tema: str = "") -> List[dict]:
|
||||
"""
|
||||
Listar archivos en el container o en un tema específico
|
||||
|
||||
|
||||
Args:
|
||||
tema: Tema/carpeta específica (opcional)
|
||||
|
||||
tema: Tema/carpeta específica (opcional) - filtrado case-insensitive
|
||||
|
||||
Returns:
|
||||
List[dict]: Lista de archivos con sus propiedades
|
||||
"""
|
||||
try:
|
||||
container_client = self.blob_service_client.get_container_client(self.container_name)
|
||||
|
||||
# Filtrar por tema si se proporciona
|
||||
name_starts_with = f"{tema}/" if tema else None
|
||||
|
||||
blobs = container_client.list_blobs(name_starts_with=name_starts_with)
|
||||
|
||||
container_client = self.blob_service_client.get_container_client(
|
||||
self.container_name
|
||||
)
|
||||
|
||||
# Obtener todos los blobs para hacer filtrado case-insensitive
|
||||
blobs = container_client.list_blobs()
|
||||
|
||||
files = []
|
||||
tema_lower = tema.lower() if tema else ""
|
||||
|
||||
for blob in blobs:
|
||||
# Extraer información del blob
|
||||
blob_tema = os.path.dirname(blob.name) if "/" in blob.name else ""
|
||||
|
||||
# Filtrar por tema de manera case-insensitive si se proporciona
|
||||
if tema and blob_tema.lower() != tema_lower:
|
||||
continue
|
||||
|
||||
blob_info = {
|
||||
"name": os.path.basename(blob.name),
|
||||
"full_path": blob.name,
|
||||
"tema": os.path.dirname(blob.name) if "/" in blob.name else "",
|
||||
"tema": blob_tema,
|
||||
"size": blob.size,
|
||||
"last_modified": blob.last_modified,
|
||||
"content_type": blob.content_settings.content_type if blob.content_settings else None
|
||||
"content_type": blob.content_settings.content_type
|
||||
if blob.content_settings
|
||||
else None,
|
||||
}
|
||||
files.append(blob_info)
|
||||
|
||||
logger.info(f"Listados {len(files)} archivos" + (f" en tema '{tema}'" if tema else ""))
|
||||
|
||||
logger.info(
|
||||
f"Listados {len(files)} archivos"
|
||||
+ (f" en tema '{tema}' (case-insensitive)" if tema else "")
|
||||
)
|
||||
return files
|
||||
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error listando archivos: {e}")
|
||||
raise e
|
||||
|
||||
|
||||
async def get_file_info(self, blob_name: str, tema: str = "") -> dict:
|
||||
"""
|
||||
Obtener información de un archivo específico
|
||||
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo
|
||||
tema: Tema/carpeta donde está el archivo
|
||||
|
||||
tema: Tema/carpeta donde está el archivo (búsqueda case-insensitive)
|
||||
|
||||
Returns:
|
||||
dict: Información del archivo
|
||||
"""
|
||||
try:
|
||||
# Construir la ruta completa
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
|
||||
# Si se proporciona tema, buscar el archivo de manera case-insensitive
|
||||
if tema:
|
||||
full_blob_name = await self._find_blob_case_insensitive(blob_name, tema)
|
||||
# Extraer el tema real del path encontrado
|
||||
real_tema = (
|
||||
os.path.dirname(full_blob_name) if "/" in full_blob_name else ""
|
||||
)
|
||||
else:
|
||||
full_blob_name = blob_name
|
||||
real_tema = ""
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
|
||||
# Obtener propiedades
|
||||
properties = blob_client.get_blob_properties()
|
||||
|
||||
|
||||
return {
|
||||
"name": blob_name,
|
||||
"full_path": full_blob_name,
|
||||
"tema": tema,
|
||||
"tema": real_tema,
|
||||
"size": properties.size,
|
||||
"last_modified": properties.last_modified,
|
||||
"content_type": properties.content_settings.content_type,
|
||||
"url": blob_client.url
|
||||
"url": blob_client.url,
|
||||
}
|
||||
|
||||
|
||||
except ResourceNotFoundError:
|
||||
logger.error(f"Archivo '{full_blob_name}' no encontrado")
|
||||
raise FileNotFoundError(f"El archivo '{blob_name}' no existe")
|
||||
except Exception as e:
|
||||
logger.error(f"Error obteniendo info del archivo '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
|
||||
async def get_download_url(self, blob_name: str, tema: str = "") -> str:
|
||||
"""
|
||||
Obtener URL de descarga directa para un archivo
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo
|
||||
tema: Tema/carpeta donde está el archivo
|
||||
tema: Tema/carpeta donde está el archivo (búsqueda case-insensitive)
|
||||
|
||||
Returns:
|
||||
str: URL de descarga
|
||||
"""
|
||||
try:
|
||||
# Construir la ruta completa
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
# Si se proporciona tema, buscar el archivo de manera case-insensitive
|
||||
if tema:
|
||||
full_blob_name = await self._find_blob_case_insensitive(blob_name, tema)
|
||||
else:
|
||||
full_blob_name = blob_name
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
return blob_client.url
|
||||
@@ -292,7 +338,9 @@ class AzureBlobService:
|
||||
logger.error(f"Error obteniendo URL de descarga para '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
async def generate_sas_url(self, blob_name: str, tema: str = "", expiry_hours: int = 1) -> str:
|
||||
async def generate_sas_url(
|
||||
self, blob_name: str, tema: str = "", expiry_hours: int = 1
|
||||
) -> str:
|
||||
"""
|
||||
Generar una URL SAS (Shared Access Signature) temporal para acceder a un archivo
|
||||
|
||||
@@ -301,7 +349,7 @@ class AzureBlobService:
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo
|
||||
tema: Tema/carpeta donde está el archivo
|
||||
tema: Tema/carpeta donde está el archivo (búsqueda case-insensitive)
|
||||
expiry_hours: Horas de validez de la URL (por defecto 1 hora)
|
||||
|
||||
Returns:
|
||||
@@ -310,13 +358,15 @@ class AzureBlobService:
|
||||
try:
|
||||
from azure.storage.blob import ContentSettings
|
||||
|
||||
# Construir la ruta completa del blob
|
||||
full_blob_name = f"{tema}/{blob_name}" if tema else blob_name
|
||||
# Si se proporciona tema, buscar el archivo de manera case-insensitive
|
||||
if tema:
|
||||
full_blob_name = await self._find_blob_case_insensitive(blob_name, tema)
|
||||
else:
|
||||
full_blob_name = blob_name
|
||||
|
||||
# Obtener cliente del blob
|
||||
blob_client = self.blob_service_client.get_blob_client(
|
||||
container=self.container_name,
|
||||
blob=full_blob_name
|
||||
container=self.container_name, blob=full_blob_name
|
||||
)
|
||||
|
||||
# Verificar que el archivo existe antes de generar el SAS
|
||||
@@ -327,11 +377,13 @@ class AzureBlobService:
|
||||
# Esto hace que el navegador muestre el PDF en lugar de descargarlo
|
||||
try:
|
||||
content_settings = ContentSettings(
|
||||
content_type='application/pdf',
|
||||
content_disposition='inline' # Clave para mostrar en navegador
|
||||
content_type="application/pdf",
|
||||
content_disposition="inline", # Clave para mostrar en navegador
|
||||
)
|
||||
blob_client.set_http_headers(content_settings=content_settings)
|
||||
logger.info(f"Headers configurados para visualización inline de '{full_blob_name}'")
|
||||
logger.info(
|
||||
f"Headers configurados para visualización inline de '{full_blob_name}'"
|
||||
)
|
||||
except Exception as e:
|
||||
logger.warning(f"No se pudieron configurar headers inline: {e}")
|
||||
|
||||
@@ -342,9 +394,9 @@ class AzureBlobService:
|
||||
# Extraer la account key del connection string para generar el SAS
|
||||
# El SAS necesita la account key para firmar el token
|
||||
account_key = None
|
||||
for part in settings.AZURE_STORAGE_CONNECTION_STRING.split(';'):
|
||||
if part.startswith('AccountKey='):
|
||||
account_key = part.split('=', 1)[1]
|
||||
for part in settings.AZURE_STORAGE_CONNECTION_STRING.split(";"):
|
||||
if part.startswith("AccountKey="):
|
||||
account_key = part.split("=", 1)[1]
|
||||
break
|
||||
|
||||
if not account_key:
|
||||
@@ -358,13 +410,15 @@ class AzureBlobService:
|
||||
account_key=account_key,
|
||||
permission=BlobSasPermissions(read=True), # Solo permisos de lectura
|
||||
expiry=expiry_time,
|
||||
start=start_time
|
||||
start=start_time,
|
||||
)
|
||||
|
||||
# Construir la URL completa con el SAS token
|
||||
sas_url = f"{blob_client.url}?{sas_token}"
|
||||
|
||||
logger.info(f"SAS URL generada para '{full_blob_name}' (válida por {expiry_hours} horas)")
|
||||
logger.info(
|
||||
f"SAS URL generada para '{full_blob_name}' (válida por {expiry_hours} horas)"
|
||||
)
|
||||
return sas_url
|
||||
|
||||
except FileNotFoundError:
|
||||
@@ -374,6 +428,47 @@ class AzureBlobService:
|
||||
logger.error(f"Error generando SAS URL para '{blob_name}': {e}")
|
||||
raise e
|
||||
|
||||
async def _find_blob_case_insensitive(self, blob_name: str, tema: str) -> str:
|
||||
"""
|
||||
Buscar un blob de manera case-insensitive
|
||||
|
||||
Args:
|
||||
blob_name: Nombre del archivo a buscar
|
||||
tema: Tema donde buscar (case-insensitive)
|
||||
|
||||
Returns:
|
||||
str: Ruta completa del blob encontrado
|
||||
|
||||
Raises:
|
||||
FileNotFoundError: Si no se encuentra el archivo
|
||||
"""
|
||||
try:
|
||||
container_client = self.blob_service_client.get_container_client(
|
||||
self.container_name
|
||||
)
|
||||
blobs = container_client.list_blobs()
|
||||
|
||||
tema_lower = tema.lower()
|
||||
blob_name_lower = blob_name.lower()
|
||||
|
||||
for blob in blobs:
|
||||
blob_tema = os.path.dirname(blob.name) if "/" in blob.name else ""
|
||||
current_blob_name = os.path.basename(blob.name)
|
||||
|
||||
if (
|
||||
blob_tema.lower() == tema_lower
|
||||
and current_blob_name.lower() == blob_name_lower
|
||||
):
|
||||
return blob.name
|
||||
|
||||
# Si no se encuentra, usar la construcción original para que falle apropiadamente
|
||||
return f"{tema}/{blob_name}"
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error buscando blob case-insensitive: {e}")
|
||||
# Fallback a construcción original
|
||||
return f"{tema}/{blob_name}"
|
||||
|
||||
|
||||
# Instancia global del servicio
|
||||
azure_service = AzureBlobService()
|
||||
azure_service = AzureBlobService()
|
||||
|
||||
@@ -28,6 +28,7 @@ dependencies = [
|
||||
# LandingAI Document AI
|
||||
"landingai-ade>=0.2.1",
|
||||
"redis-om>=0.3.5",
|
||||
"pydantic-ai-slim[google,openai]>=1.11.1",
|
||||
]
|
||||
[project.scripts]
|
||||
dev = "uvicorn app.main:app --host 0.0.0.0 --port 8000 --reload"
|
||||
|
||||
111
backend/uv.lock
generated
111
backend/uv.lock
generated
@@ -74,6 +74,7 @@ dependencies = [
|
||||
{ name = "openai" },
|
||||
{ name = "pdf2image" },
|
||||
{ name = "pillow" },
|
||||
{ name = "pydantic-ai-slim", extra = ["google", "openai"] },
|
||||
{ name = "pydantic-settings" },
|
||||
{ name = "pypdf" },
|
||||
{ name = "python-dotenv" },
|
||||
@@ -97,6 +98,7 @@ requires-dist = [
|
||||
{ name = "openai", specifier = ">=1.59.6" },
|
||||
{ name = "pdf2image", specifier = ">=1.17.0" },
|
||||
{ name = "pillow", specifier = ">=11.0.0" },
|
||||
{ name = "pydantic-ai-slim", extras = ["google", "openai"], specifier = ">=1.11.1" },
|
||||
{ name = "pydantic-settings", specifier = ">=2.10.1" },
|
||||
{ name = "pypdf", specifier = ">=5.1.0" },
|
||||
{ name = "python-dotenv", specifier = ">=1.1.1" },
|
||||
@@ -289,6 +291,19 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/e5/47/d63c60f59a59467fda0f93f46335c9d18526d7071f025cb5b89d5353ea42/fastapi-0.116.1-py3-none-any.whl", hash = "sha256:c46ac7c312df840f0c9e220f7964bada936781bc4e2e6eb71f1c4d7553786565", size = 95631, upload-time = "2025-07-11T16:22:30.485Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "genai-prices"
|
||||
version = "0.0.36"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "httpx" },
|
||||
{ name = "pydantic" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/39/e2/45c863fb61cf2d70d948e80d63e4f3db213a957976a2a3564e40ebe8f506/genai_prices-0.0.36.tar.gz", hash = "sha256:1092f5b96168967fa880440dd9dcc9287fd73910b284045f0226a38f628ccbc9", size = 46046, upload-time = "2025-11-05T14:04:13.437Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/16/89/14b4be11b74dd29827bc37b648b0540fcf3bd6530cb48031f1ce7da4594c/genai_prices-0.0.36-py3-none-any.whl", hash = "sha256:7ad39e04fbcdb5cfdc3891e68de6ca1064b6660e06e9ba76fa6f161ff12b32e4", size = 48688, upload-time = "2025-11-05T14:04:12.133Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "google-api-core"
|
||||
version = "2.28.1"
|
||||
@@ -486,6 +501,18 @@ grpc = [
|
||||
{ name = "grpcio", version = "1.76.0", source = { registry = "https://pypi.org/simple" }, marker = "python_full_version >= '3.14'" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "griffe"
|
||||
version = "1.14.0"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "colorama" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/ec/d7/6c09dd7ce4c7837e4cdb11dce980cb45ae3cd87677298dc3b781b6bce7d3/griffe-1.14.0.tar.gz", hash = "sha256:9d2a15c1eca966d68e00517de5d69dd1bc5c9f2335ef6c1775362ba5b8651a13", size = 424684, upload-time = "2025-09-05T15:02:29.167Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/2a/b1/9ff6578d789a89812ff21e4e0f80ffae20a65d5dd84e7a17873fe3b365be/griffe-1.14.0-py3-none-any.whl", hash = "sha256:0e9d52832cccf0f7188cfe585ba962d2674b241c01916d780925df34873bceb0", size = 144439, upload-time = "2025-09-05T15:02:27.511Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "grpc-google-iam-v1"
|
||||
version = "0.14.3"
|
||||
@@ -776,6 +803,18 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/76/c6/c88e154df9c4e1a2a66ccf0005a88dfb2650c1dffb6f5ce603dfbd452ce3/idna-3.10-py3-none-any.whl", hash = "sha256:946d195a0d259cbba61165e88e65941f16e9b36ea6ddb97f00452bae8b1287d3", size = 70442, upload-time = "2024-09-15T18:07:37.964Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "importlib-metadata"
|
||||
version = "8.7.0"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "zipp" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/76/66/650a33bd90f786193e4de4b3ad86ea60b53c89b669a5c7be931fac31cdb0/importlib_metadata-8.7.0.tar.gz", hash = "sha256:d13b81ad223b890aa16c5471f2ac3056cf76c5f10f82d6f9292f0b415f389000", size = 56641, upload-time = "2025-04-27T15:29:01.736Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/20/b0/36bd937216ec521246249be3bf9855081de4c5e06a0c9b4219dbeda50373/importlib_metadata-8.7.0-py3-none-any.whl", hash = "sha256:e5dd1551894c77868a30651cef00984d50e1002d06942a7101d34870c5f02afd", size = 27656, upload-time = "2025-04-27T15:29:00.214Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "isodate"
|
||||
version = "0.7.2"
|
||||
@@ -1009,6 +1048,15 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/98/4c/6c0c338ca7182e4ecb7af61049415e7b3513cc6cea9aa5bf8ca508f53539/langsmith-0.4.41-py3-none-any.whl", hash = "sha256:5cdc554e5f0361bf791fdd5e8dea16d5ba9dfce09b3b8f8bba5e99450c569b27", size = 399279, upload-time = "2025-11-04T22:31:30.268Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "logfire-api"
|
||||
version = "4.14.2"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/59/25/6072086af3b3ac5c2c2f2a6cf89488a1b228ffc6ee0fb357ed1e227efd13/logfire_api-4.14.2.tar.gz", hash = "sha256:bbdeccd931069b76ab811261b41bc52d8b78d1c045fc4b4237dbc085e0fb9bcd", size = 57604, upload-time = "2025-10-24T20:14:40.551Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/58/c7/b06a83df678fca882c24fb498e628e0406bdb95ffdfa7ae43ecc0a714d52/logfire_api-4.14.2-py3-none-any.whl", hash = "sha256:aa4af2ecb007c3e0095e25ba4526fd8c0e2c0be2ceceac71ca651c4ad86dc713", size = 95021, upload-time = "2025-10-24T20:14:36.161Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "more-itertools"
|
||||
version = "10.8.0"
|
||||
@@ -1100,6 +1148,19 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/8c/74/6bfc3adc81f6c2cea4439f2a734c40e3a420703bbcdc539890096a732bbd/openai-2.7.1-py3-none-any.whl", hash = "sha256:2f2530354d94c59c614645a4662b9dab0a5b881c5cd767a8587398feac0c9021", size = 1008780, upload-time = "2025-11-04T06:07:20.818Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "opentelemetry-api"
|
||||
version = "1.38.0"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "importlib-metadata" },
|
||||
{ name = "typing-extensions" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/08/d8/0f354c375628e048bd0570645b310797299754730079853095bf000fba69/opentelemetry_api-1.38.0.tar.gz", hash = "sha256:f4c193b5e8acb0912b06ac5b16321908dd0843d75049c091487322284a3eea12", size = 65242, upload-time = "2025-10-16T08:35:50.25Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/ae/a2/d86e01c28300bd41bab8f18afd613676e2bd63515417b77636fc1add426f/opentelemetry_api-1.38.0-py3-none-any.whl", hash = "sha256:2891b0197f47124454ab9f0cf58f3be33faca394457ac3e09daba13ff50aa582", size = 65947, upload-time = "2025-10-16T08:35:30.23Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "orjson"
|
||||
version = "3.11.4"
|
||||
@@ -1364,6 +1425,32 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/6a/c0/ec2b1c8712ca690e5d61979dee872603e92b8a32f94cc1b72d53beab008a/pydantic-2.11.7-py3-none-any.whl", hash = "sha256:dde5df002701f6de26248661f6835bbe296a47bf73990135c7d07ce741b9623b", size = 444782, upload-time = "2025-06-14T08:33:14.905Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pydantic-ai-slim"
|
||||
version = "1.11.1"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "genai-prices" },
|
||||
{ name = "griffe" },
|
||||
{ name = "httpx" },
|
||||
{ name = "opentelemetry-api" },
|
||||
{ name = "pydantic" },
|
||||
{ name = "pydantic-graph" },
|
||||
{ name = "typing-inspection" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/90/a5/fbfcdd3c89549dd44417606af0130f1118aea8e43f4d14723e49218901a6/pydantic_ai_slim-1.11.1.tar.gz", hash = "sha256:242fb5c7a0f812d540f68d4e2e6498730ef11644b55ccf3da38bf9767802f742", size = 298765, upload-time = "2025-11-06T00:48:42.815Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/f8/6d/d8ea48afdd8838d6419cdbc08d81753e2e732ff3451e3d83f6b4b56388af/pydantic_ai_slim-1.11.1-py3-none-any.whl", hash = "sha256:00ca8b0a8f677fa9efd077239b66c925423d1dc517dfac7953b62547a66adbf2", size = 397971, upload-time = "2025-11-06T00:48:28.219Z" },
|
||||
]
|
||||
|
||||
[package.optional-dependencies]
|
||||
google = [
|
||||
{ name = "google-genai" },
|
||||
]
|
||||
openai = [
|
||||
{ name = "openai" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pydantic-core"
|
||||
version = "2.33.2"
|
||||
@@ -1406,6 +1493,21 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/6f/9a/e73262f6c6656262b5fdd723ad90f518f579b7bc8622e43a942eec53c938/pydantic_core-2.33.2-cp313-cp313t-win_amd64.whl", hash = "sha256:c2fc0a768ef76c15ab9238afa6da7f69895bb5d1ee83aeea2e3509af4472d0b9", size = 1935777, upload-time = "2025-04-23T18:32:25.088Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pydantic-graph"
|
||||
version = "1.11.1"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
dependencies = [
|
||||
{ name = "httpx" },
|
||||
{ name = "logfire-api" },
|
||||
{ name = "pydantic" },
|
||||
{ name = "typing-inspection" },
|
||||
]
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/6f/b6/1b37a9517bc71fde33184cc6f3f03795c3669b7be5a143a3012fb112742d/pydantic_graph-1.11.1.tar.gz", hash = "sha256:345d6309ac677ef6cf2f5b225e6762afd9b87cc916b943376a5cb555705a7f2b", size = 57964, upload-time = "2025-11-06T00:48:45.028Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/50/64/934e1f9be64f44515c501bf528cfc2dd672516530d5a7aa7436f72aba5ef/pydantic_graph-1.11.1-py3-none-any.whl", hash = "sha256:4d52d0c925672439e407d64e663a5e7f011f0bb0941c8b6476911044c7478cd6", size = 72002, upload-time = "2025-11-06T00:48:32.411Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pydantic-settings"
|
||||
version = "2.10.1"
|
||||
@@ -2148,6 +2250,15 @@ wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/0f/c9/7243eb3f9eaabd1a88a5a5acadf06df2d83b100c62684b7425c6a11bcaa8/xxhash-3.6.0-cp314-cp314t-win_arm64.whl", hash = "sha256:bb79b1e63f6fd84ec778a4b1916dfe0a7c3fdb986c06addd5db3a0d413819d95", size = 28898, upload-time = "2025-10-02T14:36:17.843Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zipp"
|
||||
version = "3.23.0"
|
||||
source = { registry = "https://pypi.org/simple" }
|
||||
sdist = { url = "https://files.pythonhosted.org/packages/e3/02/0f2892c661036d50ede074e376733dca2ae7c6eb617489437771209d4180/zipp-3.23.0.tar.gz", hash = "sha256:a07157588a12518c9d4034df3fbbee09c814741a33ff63c05fa29d26a2404166", size = 25547, upload-time = "2025-06-08T17:06:39.4Z" }
|
||||
wheels = [
|
||||
{ url = "https://files.pythonhosted.org/packages/2e/54/647ade08bf0db230bfea292f893923872fd20be6ac6f53b2b936ba839d75/zipp-3.23.0-py3-none-any.whl", hash = "sha256:071652d6115ed432f5ce1d34c336c0adfd6a884660d1e9712a256d3d3bd4b14e", size = 10276, upload-time = "2025-06-08T17:06:38.034Z" },
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zstandard"
|
||||
version = "0.25.0"
|
||||
|
||||
Reference in New Issue
Block a user