import os import gradio as gr import aiohttp import asyncio import json from datasets import Dataset, DatasetDict, load_dataset from huggingface_hub import HfFolder # 從環境變量中獲取 Hugging Face API 令牌和其他配置 HF_API_TOKEN = os.environ.get("Feedback_API_TOKEN") LLM_API = os.environ.get("LLM_API") LLM_URL = os.environ.get("LLM_URL") USER_ID = "HuggingFace Space" DATASET_NAME = os.environ.get("DATASET_NAME") # 確保令牌不為空 if HF_API_TOKEN is None: raise ValueError("HF_API_TOKEN 環境變量未設置。請在 Hugging Face Space 的設置中添加該環境變量。") # 設置 Hugging Face API 令牌 HfFolder.save_token(HF_API_TOKEN) # 定義數據集特徵 features = { "user_input": "string", "response": "string", "feedback_type": "string", "improvement": "string" } # 加載或創建數據集 try: dataset = load_dataset(DATASET_NAME) except: dataset = DatasetDict({ "feedback": Dataset.from_dict({ "user_input": [], "response": [], "feedback_type": [], "improvement": [] }) }) async def send_chat_message(LLM_URL, LLM_API, user_input): payload = { "inputs": {}, "query": user_input, "response_mode": "streaming", "conversation_id": "", "user": USER_ID, } print("Sending chat message payload:", payload) # Debug information async with aiohttp.ClientSession() as session: try: async with session.post( url=f"{LLM_URL}/chat-messages", headers={"Authorization": f"Bearer {LLM_API}"}, json=payload, timeout=aiohttp.ClientTimeout(total=60) ) as response: if response.status != 200: print(f"Error: {response.status}") return f"Error: {response.status}" full_response = [] async for line in response.content: line = line.decode('utf-8').strip() if not line: continue if "data: " not in line: continue try: print("Received line:", line) # Debug information data = json.loads(line.split("data: ")[1]) if "answer" in data: full_response.append(data["answer"]) except (IndexError, json.JSONDecodeError) as e: print(f"Error parsing line: {line}, error: {e}") # Debug information continue if full_response: return ''.join(full_response).strip() else: return "Error: No response found in the response" except Exception as e: print(f"Exception: {e}") return f"Exception: {e}" async def handle_input_async(user_input): print(f"Handling input: {user_input}") chat_response = await send_chat_message(LLM_URL, LLM_API, user_input) print("Chat response:", chat_response) # Debug information return chat_response def handle_input(user_input): print(f"Handling input synchronously: {user_input}") return asyncio.run(handle_input_async(user_input)) def save_feedback(user_input, response, feedback_type, improvement): feedback = { "user_input": user_input, "response": response, "feedback_type": feedback_type, "improvement": improvement } print(f"Saving feedback: {feedback}") # Append to the dataset new_data = { "user_input": [user_input], "response": [response], "feedback_type": [feedback_type], "improvement": [improvement] } global dataset dataset["feedback"] = Dataset.from_dict({ "user_input": dataset["feedback"]["user_input"] + [user_input], "response": dataset["feedback"]["response"] + [response], "feedback_type": dataset["feedback"]["feedback_type"] + [feedback_type], "improvement": dataset["feedback"]["improvement"] + [improvement] }) dataset.push_to_hub(DATASET_NAME) def handle_feedback(response, feedback_type, improvement): global last_user_input save_feedback(last_user_input, response, feedback_type, improvement) return "感謝您的反饋!" # 讀取並顯示反饋內容的函數 def show_feedback(): try: feedbacks = dataset["feedback"].to_pandas().to_dict(orient="records") print(f"Feedbacks: {feedbacks}") # Debug information return feedbacks except Exception as e: print(f"Error: {e}") # Debug information return {"error": str(e)} TITLE = """