Spaces:
Build error
Build error
Update app.py to include additional print statement for debugging the chat function and return the full result from the model.
3b89341
| from fastapi import FastAPI | |
| from pydantic import BaseModel | |
| from llama_cpp import Llama | |
| from pathlib import Path | |
| import os | |
| from dotenv import load_dotenv | |
| from huggingface_hub import hf_hub_download | |
| # Carica variabili da .env | |
| load_dotenv() | |
| HUGGINGFACE_HUB_TOKEN = os.environ.get("HUGGINGFACE_HUB_TOKEN") | |
| app = FastAPI(title="Company Assistant API") | |
| # Percorso del modello locale | |
| MODEL_DIR = Path("models/llama-3.2-1b") | |
| MODEL_DIR.mkdir(parents=True, exist_ok=True) | |
| MODEL_PATH = MODEL_DIR / "llama-3.2-1b-instruct-q4_k_m.gguf" | |
| print('ciao') | |
| # Scarica il modello se non esiste | |
| if not MODEL_PATH.exists(): | |
| print("Scaricando il modello GGUF Llama 3.2-1B...") | |
| MODEL_PATH = hf_hub_download( | |
| repo_id="hugging-quants/Llama-3.2-1B-Instruct-Q4_K_M-GGUF", | |
| filename="llama-3.2-1b-instruct-q4_k_m.gguf", | |
| cache_dir=str(MODEL_DIR), | |
| token=HUGGINGFACE_HUB_TOKEN | |
| ) | |
| print("Modello scaricato!") | |
| # Inizializza il modello | |
| llm = Llama(model_path=str(MODEL_PATH)) | |
| # Modello dati richiesta | |
| class Message(BaseModel): | |
| text: str | |
| def root(): | |
| return {"message": "Company Assistant API è attiva!"} | |
| def chat(message: Message): | |
| print(message) | |
| result = llm(prompt=message.text, max_tokens=200) | |
| print('risultato ', result) | |
| return {"reply": result} | |