gguf-convertor / app.py
zackli4ai's picture
Update app.py (#2)
584d84d verified
import gradio as gr
from gradio_huggingfacehub_search import HuggingfaceHubSearch
import requests
processed_inputs = {}
def process_inputs(model_id, q_method, email, oauth_token: gr.OAuthToken | None, profile: gr.OAuthProfile | None):
if oauth_token is None or profile is None or oauth_token.token is None or profile.username is None:
return "### You must be logged in to use this service."
if not model_id or not q_method or not email:
return "### All fields are required!"
input_hash = hash((model_id, q_method, oauth_token.token, profile.username))
if input_hash in processed_inputs and processed_inputs[input_hash] == 200:
return "### Oops! 😲 Looks like you've already submitted this task 🚀. Please hang tight! We'll send you an email with all the details once it's ready 💌. Thanks for your patience! 😊"
url = "https://sdk.nexa4ai.com/task"
data = {
"repository_url": f"https://huggingface.co/{model_id}",
"username": profile.username,
"access_token": oauth_token.token,
"email": email,
"quantization_option": q_method,
}
"""
# OAuth Token Information:
# - This is an OAuth token, not a user's password.
# - We need the OAuth token to clone the related repository and access its contents.
# - As mentioned in the README.md, only read permission is requested, which includes:
# - Read access to your public profile
# - Read access to the content of all your public repos
# - The token expires after 60 minutes.
# - For more information about OAuth, please refer to the official documentation:
# https://huggingface.co/docs/hub/en/spaces-oauth
"""
response = requests.post(url, json=data)
if response.status_code == 200:
processed_inputs[input_hash] = 200
return "### Your request has been submitted successfully! 🌟 We'll notify you by email 📧 once everything is processed. 😊"
else:
processed_inputs[input_hash] = response.status_code
return f"### Failed to submit request: {response.text}"
iface = gr.Interface(
fn=process_inputs,
inputs=[
HuggingfaceHubSearch(
label="Hub Model ID",
placeholder="Search for model id on Huggingface",
search_type="model",
),
gr.Dropdown(
["q2_K", "q3_K", "q3_K_S", "q3_K_M", "q3_K_L", "q4_0", "q4_1", "q4_K", "q4_K_S", "q4_K_M", "q5_0", "q5_1", "q5_K", "q5_K_S", "q5_K_M", "q6_K", "q8_0", "f16"],
label="Quantization Option",
info="GGML quantisation options",
value="q4_0",
filterable=False
),
gr.Textbox(label="Email", placeholder="Enter your email here")
],
outputs = gr.Markdown(
label="output",
value="### Please enter the model URL, select a quantization method, and provide your email address."
),
allow_flagging="never"
)
theme = gr.themes.Soft(text_size="lg", spacing_size="lg")
with gr.Blocks(theme=theme) as demo:
with gr.Row(variant="panel'"):
gr.Markdown(value="## 🚀 Unleash the Power of Custom GGML Quantized Models! ⚡"),
gr.LoginButton(min_width=380)
gr.Markdown(value="🚨 **IMPORTANT:** You **MUST** grant access to the model repository before use.")
gr.Markdown(value="🔔 You **MUST** be logged in to use this service.")
iface.render()
gr.Markdown(value="We sincerely thank our community members, [Perry](https://huggingface.co/PerryCheng614), [Brian](https://huggingface.co/JoyboyBrian), [Qi](https://huggingface.co/qiqiWav), [David](https://huggingface.co/Davidqian123), for their extraordinary contributions to this GGUF converter project.")
demo.launch(share=True)