Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
import gradio as gr | |
from gradio_huggingfacehub_search import HuggingfaceHubSearch | |
import requests | |
processed_inputs = {} | |
def process_inputs(model_id, q_method, email, oauth_token: gr.OAuthToken | None, profile: gr.OAuthProfile | None): | |
if oauth_token is None or profile is None or oauth_token.token is None or profile.username is None: | |
return "### You must be logged in to use this service." | |
if not model_id or not q_method or not email: | |
return "### All fields are required!" | |
input_hash = hash((model_id, q_method, oauth_token.token, profile.username)) | |
if input_hash in processed_inputs and processed_inputs[input_hash] == 200: | |
return "### Oops! 😲 Looks like you've already submitted this task 🚀. Please hang tight! We'll send you an email with all the details once it's ready 💌. Thanks for your patience! 😊" | |
url = "https://sdk.nexa4ai.com/task" | |
data = { | |
"repository_url": f"https://huggingface.co/{model_id}", | |
"username": profile.username, | |
"access_token": oauth_token.token, | |
"email": email, | |
"quantization_option": q_method, | |
} | |
""" | |
# OAuth Token Information: | |
# - This is an OAuth token, not a user's password. | |
# - We need the OAuth token to clone the related repository and access its contents. | |
# - As mentioned in the README.md, only read permission is requested, which includes: | |
# - Read access to your public profile | |
# - Read access to the content of all your public repos | |
# - The token expires after 60 minutes. | |
# - For more information about OAuth, please refer to the official documentation: | |
# https://huggingface.co/docs/hub/en/spaces-oauth | |
""" | |
response = requests.post(url, json=data) | |
if response.status_code == 200: | |
processed_inputs[input_hash] = 200 | |
return "### Your request has been submitted successfully! 🌟 We'll notify you by email 📧 once everything is processed. 😊" | |
else: | |
processed_inputs[input_hash] = response.status_code | |
return f"### Failed to submit request: {response.text}" | |
iface = gr.Interface( | |
fn=process_inputs, | |
inputs=[ | |
HuggingfaceHubSearch( | |
label="Hub Model ID", | |
placeholder="Search for model id on Huggingface", | |
search_type="model", | |
), | |
gr.Dropdown( | |
["q2_K", "q3_K", "q3_K_S", "q3_K_M", "q3_K_L", "q4_0", "q4_1", "q4_K", "q4_K_S", "q4_K_M", "q5_0", "q5_1", "q5_K", "q5_K_S", "q5_K_M", "q6_K", "q8_0", "f16"], | |
label="Quantization Option", | |
info="GGML quantisation options", | |
value="q4_0", | |
filterable=False | |
), | |
gr.Textbox(label="Email", placeholder="Enter your email here") | |
], | |
outputs = gr.Markdown( | |
label="output", | |
value="### Please enter the model URL, select a quantization method, and provide your email address." | |
), | |
allow_flagging="never" | |
) | |
theme = gr.themes.Soft(text_size="lg", spacing_size="lg") | |
with gr.Blocks(theme=theme) as demo: | |
with gr.Row(variant="panel'"): | |
gr.Markdown(value="## 🚀 Unleash the Power of Custom GGML Quantized Models! ⚡"), | |
gr.LoginButton(min_width=380) | |
gr.Markdown(value="🚨 **IMPORTANT:** You **MUST** grant access to the model repository before use.") | |
gr.Markdown(value="🔔 You **MUST** be logged in to use this service.") | |
iface.render() | |
gr.Markdown(value="We sincerely thank our community members, [Perry](https://huggingface.co/PerryCheng614), [Brian](https://huggingface.co/JoyboyBrian), [Qi](https://huggingface.co/qiqiWav), [David](https://huggingface.co/Davidqian123), for their extraordinary contributions to this GGUF converter project.") | |
demo.launch(share=True) |