| import gradio as gr |
| import openai |
| import os |
| from typing import Optional |
| from gradio_client import Client |
|
|
| |
|
|
| |
| openai.api_key = os.getenv("OPENAI_API_KEY") |
| if not openai.api_key: |
| raise ValueError("OpenAI API ํ ํฐ(OPENAI_API_KEY)์ด ์ค์ ๋์ง ์์์ต๋๋ค.") |
|
|
| def call_openai_api( |
| content: str, |
| system_message: str, |
| max_tokens: int, |
| temperature: float, |
| top_p: float |
| ) -> str: |
| """ |
| OpenAI์ GPT-4o-mini ๋ชจ๋ธ์ ์ด์ฉํด ํ ๋ฒ์ ์ง๋ฌธ(content)์ ๋ํ ๋ต๋ณ์ ๋ฐํํ๋ ํจ์. |
| """ |
| try: |
| response = openai.ChatCompletion.create( |
| model="gpt-4o-mini", |
| messages=[ |
| {"role": "system", "content": system_message}, |
| {"role": "user", "content": content}, |
| ], |
| max_tokens=max_tokens, |
| temperature=temperature, |
| top_p=top_p, |
| ) |
| assistant_message = response.choices[0].message['content'] |
| return assistant_message |
| except Exception as e: |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" |
|
|
| |
| |
| |
|
|
| OPENAI_SYSTEM_MESSAGE = """๋ฐ๋์ ํ๊ธ๋ก ๋ต๋ณํ ๊ฒ. |
| ๋๋ ์ต๊ณ ์ ๋น์์ด๋ค. |
| ๋ด๊ฐ ์๊ตฌํ๋ ๊ฒ๋ค์ ์ต๋ํ ์์ธํ๊ณ ์ ํํ๊ฒ ๋ต๋ณํ๋ผ. |
| ##[๊ธฐ๋ณธ๊ท์น] |
| 1. ๋ฐ๋์ ํ๊ตญ์ด(ํ๊ธ)๋ก ์์ฑํ๋ผ. |
| 2. ๋๋ ๊ฐ์ฅ ์ฃผ๋ชฉ๋ฐ๋ ๋ง์ผํฐ์ด๋ฉฐ ๋ธ๋ก๊ทธ ๋ง์ผํ
์ ๋ฌธ๊ฐ์ด๋ค. |
| 3. ํนํ ๋๋ '์ ๋ณด์ฑ(Informative)' ์ ๋ฌธ ๋ธ๋ก๊ทธ ๋ง์ผํ
์ ๋ฌธ๊ฐ์ด๋ค. |
| 4. ์ ๋ณด ์ ๊ณต์ ์ด์ ์ ๋ง์ถ์ด ์์ฑํ๋ค. |
| ##[ํ
์คํธ ์์ฑ ๊ท์น] |
| 1. ์์ฃผ์ ๋ฅผ 5๊ฐ๋ก ๊ตฌ๋ถํ์ฌ 2000์ ์ด์๋๋๋ก ์์ฑํ๋ผ. |
| 2. ์ ์ฒด ๋งฅ๋ฝ์ ์ดํดํ๊ณ ๋ฌธ์ฅ์ ์ผ๊ด์ฑ์ ์ ์งํ๋ผ. |
| 3. ์ ๋๋ก ์ฐธ๊ณ ๊ธ์ ํ๋ฌธ์ฅ ์ด์ ๊ทธ๋๋ก ์ถ๋ ฅํ์ง ๋ง ๊ฒ. |
| 4. ์ฃผ์ ์ ์ํฉ์ ๋ง๋ ์ ์ ํ ์ดํ๋ฅผ ์ ํํ๋ผ. |
| 5. ํ๊ธ ์ดํ์ ๋์ด๋๋ ์ฝ๊ฒ ์์ฑํ๋ผ. |
| 6. ์ ๋ ๋ฌธ์ฅ์ ๋์ '๋ต๋๋ค'๋ฅผ ์ฌ์ฉํ์ง ๋ง ๊ฒ. |
| ###[์ ๋ณด์ฑ ๋ธ๋ก๊ทธ ์์ฑ ๊ท์น] |
| 1. ๋
์๊ฐ ์ป๊ณ ์ ํ๋ ์ ์ฉํ ์ ๋ณด์ ํฅ๋ฏธ๋ก์ด ์ ๋ณด๋ฅผ ์ ๊ณตํ๋๋ก ์์ฑํ๋ผ. |
| 2. ๋
์์ ๊ณต๊ฐ์ ์ด๋์ด๋ด๊ณ ๊ถ๊ธ์ฆ์ ํด๊ฒฐํ๋๋ก ์์ฑํ๋ผ. |
| 3. ๋
์์ ๊ด์ฌ์ฌ๋ฅผ ์ถฉ์กฑ์ํค๋๋ก ์์ฑํ๋ผ. |
| 4. ๋
์์๊ฒ ์ด๋์ด ๋๋ ์ ๋ณด๋ฅผ ์์ฑํ๋ผ. |
| ##[์ ์ธ ๊ท์น] |
| 1. ๋ฐ๋์ ๋น์์ด ๋ฐ ์์ค(expletive, abusive language, slang)์ ์ ์ธํ๋ผ. |
| 2. ๋ฐ๋์ ์ฐธ๊ณ ๊ธ์ ๋งํฌ(URL)๋ ์ ์ธํ๋ผ. |
| 3. ์ฐธ๊ณ ๊ธ์์ '๋งํฌ๋ฅผ ํ์ธํด์ฃผ์ธ์'์ ๊ฐ์ ๋งํฌ ์ด๋์ ๋ฌธ๊ตฌ๋ ์ ์ธํ๋ผ. |
| 4. ์ฐธ๊ณ ๊ธ์ ์๋ ์์ฑ์, ํ์, ์ ํ๋ฒ, ๊ธฐ์์ ์ด๋ฆ, ์ ์นญ, ๋๋ค์์ ๋ฐ๋์ ์ ์ธํ๋ผ. |
| 5. ๋ฐ๋์ ๋ฌธ์ฅ์ ๋๋ถ๋ถ์ด ์ด์ํ ํ๊ตญ์ด ํํ์ ์ ์ธํ๋ผ('์์', '๋ต๋๋ค', 'ํด์', 'ํด์ฃผ์ฃ ', '๋์ฃ ', '๋์ด์', '๊ณ ์' ๋ฑ.) |
| """ |
|
|
| OPENAI_MAX_TOKENS = 4000 |
| OPENAI_TEMPERATURE = 0.7 |
| OPENAI_TOP_P = 0.95 |
|
|
| |
| |
| |
| client = Client("Kims12/blog") |
|
|
| |
| |
| |
| with gr.Blocks() as demo: |
| gr.Markdown("# ๋ธ๋ก๊ทธ ์์ฑ๊ธฐ") |
|
|
| |
| tone_radio = gr.Radio( |
| label="๋งํฌ๋ฐ๊พธ๊ธฐ", |
| choices=["์น๊ทผํ๊ฒ", "์ผ๋ฐ์ ์ธ", "์ ๋ฌธ์ ์ธ"], |
| value="์ผ๋ฐ์ ์ธ" |
| ) |
|
|
| |
| with gr.Row(): |
| with gr.Column(): |
| ref1 = gr.Textbox(label="์ฐธ์กฐ๊ธ 1") |
| fetch_button1 = gr.Button("API ์คํ1") |
| with gr.Column(): |
| ref2 = gr.Textbox(label="์ฐธ์กฐ๊ธ 2") |
| fetch_button2 = gr.Button("API ์คํ2") |
| with gr.Column(): |
| ref3 = gr.Textbox(label="์ฐธ์กฐ๊ธ 3") |
| fetch_button3 = gr.Button("API ์คํ3") |
|
|
| output_box = gr.Textbox(label="๊ฒฐ๊ณผ", lines=20, interactive=False) |
|
|
| |
| def fetch_ref1(url: str) -> str: |
| if not url: |
| return "URL์ ์
๋ ฅํด์ฃผ์ธ์." |
| try: |
| result = client.predict( |
| url=url, |
| api_name="/predict" |
| ) |
| return result |
| except Exception as e: |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" |
|
|
| |
| def fetch_ref2(url: str) -> str: |
| if not url: |
| return "URL์ ์
๋ ฅํด์ฃผ์ธ์." |
| try: |
| result = client.predict( |
| url=url, |
| api_name="/predict" |
| ) |
| return result |
| except Exception as e: |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" |
|
|
| |
| def fetch_ref3(url: str) -> str: |
| if not url: |
| return "URL์ ์
๋ ฅํด์ฃผ์ธ์." |
| try: |
| result = client.predict( |
| url=url, |
| api_name="/predict" |
| ) |
| return result |
| except Exception as e: |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" |
|
|
| |
| fetch_button1.click( |
| fn=fetch_ref1, |
| inputs=ref1, |
| outputs=ref1 |
| ) |
|
|
| fetch_button2.click( |
| fn=fetch_ref2, |
| inputs=ref2, |
| outputs=ref2 |
| ) |
|
|
| fetch_button3.click( |
| fn=fetch_ref3, |
| inputs=ref3, |
| outputs=ref3 |
| ) |
|
|
| |
| def generate_blog(tone_value: str, ref1_value: str, ref2_value: str, ref3_value: str) -> str: |
| |
| question = ( |
| f"๋งํฌ: {tone_value}\n" |
| f"์ฐธ์กฐ๊ธ1: {ref1_value}\n" |
| f"์ฐธ์กฐ๊ธ2: {ref2_value}\n" |
| f"์ฐธ์กฐ๊ธ3: {ref3_value}\n" |
| ) |
|
|
| |
| response = call_openai_api( |
| content=question, |
| system_message=OPENAI_SYSTEM_MESSAGE, |
| max_tokens=OPENAI_MAX_TOKENS, |
| temperature=OPENAI_TEMPERATURE, |
| top_p=OPENAI_TOP_P |
| ) |
| return response |
|
|
| generate_button = gr.Button("์์ฑํ๊ธฐ") |
| generate_button.click( |
| fn=generate_blog, |
| inputs=[tone_radio, ref1, ref2, ref3], |
| outputs=output_box |
| ) |
|
|
| if __name__ == "__main__": |
| demo.launch() |
|
|