Spaces:
Sleeping
Sleeping
Update app.py
Browse filesadded show_api
app.py
CHANGED
@@ -13,13 +13,11 @@ inferenceModel = inference_model.to(device)
|
|
13 |
|
14 |
def generate():
|
15 |
context = torch.zeros((1, 1), dtype=torch.long, device=device)
|
16 |
-
print("Context prepared")
|
17 |
output = data_utils.decode(inferenceModel.generate(context, max_new_tokens=500)[0].tolist())
|
18 |
-
print("Output : ", output)
|
19 |
return output
|
20 |
|
21 |
|
22 |
demo = gr.Interface(fn=generate, inputs=None, outputs="text", title="F.R.I.E.N.D.S GPT", thumbnail="FRIENDS.jpg")
|
23 |
|
24 |
if __name__ == "__main__":
|
25 |
-
demo.launch(share=True)
|
|
|
13 |
|
14 |
def generate():
|
15 |
context = torch.zeros((1, 1), dtype=torch.long, device=device)
|
|
|
16 |
output = data_utils.decode(inferenceModel.generate(context, max_new_tokens=500)[0].tolist())
|
|
|
17 |
return output
|
18 |
|
19 |
|
20 |
demo = gr.Interface(fn=generate, inputs=None, outputs="text", title="F.R.I.E.N.D.S GPT", thumbnail="FRIENDS.jpg")
|
21 |
|
22 |
if __name__ == "__main__":
|
23 |
+
demo.launch(show_api=False, share=True)
|