Spaces:
Running
Running
| """ | |
| Main FastAPI application for text summarizer backend. | |
| """ | |
| import time | |
| from fastapi import FastAPI | |
| from fastapi.middleware.cors import CORSMiddleware | |
| from app.core.config import settings | |
| from app.core.logging import setup_logging, get_logger | |
| from app.api.v1.routes import api_router | |
| from app.api.v2.routes import api_router as v2_api_router | |
| from app.core.middleware import request_context_middleware | |
| from app.core.errors import init_exception_handlers | |
| from app.services.summarizer import ollama_service | |
| from app.services.transformers_summarizer import transformers_service | |
| from app.services.hf_streaming_summarizer import hf_streaming_service | |
| # Set up logging | |
| setup_logging() | |
| logger = get_logger(__name__) | |
| # Create FastAPI app | |
| app = FastAPI( | |
| title="Text Summarizer API", | |
| description="A FastAPI backend with multiple summarization engines: V1 (Ollama + Transformers pipeline) and V2 (HuggingFace streaming)", | |
| version="2.0.0", | |
| docs_url="/docs", | |
| redoc_url="/redoc", | |
| ) | |
| # Add CORS middleware | |
| app.add_middleware( | |
| CORSMiddleware, | |
| allow_origins=["*"], # Configure appropriately for production | |
| allow_credentials=True, | |
| allow_methods=["*"], | |
| allow_headers=["*"], | |
| ) | |
| # Add request context middleware | |
| app.middleware("http")(request_context_middleware) | |
| # Initialize exception handlers | |
| init_exception_handlers(app) | |
| # Include API routes | |
| app.include_router(api_router, prefix="/api/v1") | |
| app.include_router(v2_api_router, prefix="/api/v2") | |
| async def startup_event(): | |
| """Application startup event.""" | |
| logger.info("Starting Text Summarizer API") | |
| logger.info(f"V1 warmup enabled: {settings.enable_v1_warmup}") | |
| logger.info(f"V2 warmup enabled: {settings.enable_v2_warmup}") | |
| # V1 Ollama warmup (conditional) | |
| if settings.enable_v1_warmup: | |
| logger.info(f"Ollama host: {settings.ollama_host}") | |
| logger.info(f"Ollama model: {settings.ollama_model}") | |
| # Validate Ollama connectivity | |
| try: | |
| is_healthy = await ollama_service.check_health() | |
| if is_healthy: | |
| logger.info("β Ollama service is accessible and healthy") | |
| else: | |
| logger.warning("β οΈ Ollama service is not responding properly") | |
| logger.warning(f" Please ensure Ollama is running at {settings.ollama_host}") | |
| logger.warning(f" And that model '{settings.ollama_model}' is available") | |
| except Exception as e: | |
| logger.error(f"β Failed to connect to Ollama: {e}") | |
| logger.error(f" Please check that Ollama is running at {settings.ollama_host}") | |
| logger.error(f" And that model '{settings.ollama_model}' is installed") | |
| # Warm up the Ollama model | |
| logger.info("π₯ Warming up Ollama model...") | |
| try: | |
| warmup_start = time.time() | |
| await ollama_service.warm_up_model() | |
| warmup_time = time.time() - warmup_start | |
| logger.info(f"β Ollama model warmup completed in {warmup_time:.2f}s") | |
| except Exception as e: | |
| logger.warning(f"β οΈ Ollama model warmup failed: {e}") | |
| else: | |
| logger.info("βοΈ Skipping V1 Ollama warmup (disabled)") | |
| # V1 Transformers pipeline warmup (always enabled for backward compatibility) | |
| logger.info("π₯ Warming up Transformers pipeline model...") | |
| try: | |
| pipeline_start = time.time() | |
| await transformers_service.warm_up_model() | |
| pipeline_time = time.time() - pipeline_start | |
| logger.info(f"β Pipeline warmup completed in {pipeline_time:.2f}s") | |
| except Exception as e: | |
| logger.warning(f"β οΈ Pipeline warmup failed: {e}") | |
| # V2 HuggingFace warmup (conditional) | |
| if settings.enable_v2_warmup: | |
| logger.info(f"HuggingFace model: {settings.hf_model_id}") | |
| logger.info("π₯ Warming up HuggingFace model...") | |
| try: | |
| hf_start = time.time() | |
| await hf_streaming_service.warm_up_model() | |
| hf_time = time.time() - hf_start | |
| logger.info(f"β HuggingFace model warmup completed in {hf_time:.2f}s") | |
| except Exception as e: | |
| logger.warning(f"β οΈ HuggingFace model warmup failed: {e}") | |
| else: | |
| logger.info("βοΈ Skipping V2 HuggingFace warmup (disabled)") | |
| async def shutdown_event(): | |
| """Application shutdown event.""" | |
| logger.info("Shutting down Text Summarizer API") | |
| async def root(): | |
| """Root endpoint.""" | |
| return { | |
| "message": "Text Summarizer API", | |
| "version": "1.0.0", | |
| "docs": "/docs" | |
| } | |
| async def health_check(): | |
| """Health check endpoint.""" | |
| return { | |
| "status": "ok", | |
| "service": "text-summarizer-api", | |
| "version": "1.0.0" | |
| } | |
| async def debug_config(): | |
| """Debug endpoint to show current configuration.""" | |
| return { | |
| "ollama_host": settings.ollama_host, | |
| "ollama_model": settings.ollama_model, | |
| "ollama_timeout": settings.ollama_timeout, | |
| "server_host": settings.server_host, | |
| "server_port": settings.server_port, | |
| "hf_model_id": settings.hf_model_id, | |
| "hf_device_map": settings.hf_device_map, | |
| "enable_v1_warmup": settings.enable_v1_warmup, | |
| "enable_v2_warmup": settings.enable_v2_warmup | |
| } | |