rushi-k commited on
Commit
98b0729
1 Parent(s): 2f07c4b

Create inference.py

Browse files
Files changed (1) hide show
  1. inference.py +28 -0
inference.py ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from transformers import GPT2Tokenizer, GPT2LMHeadModel
3
+ from fastapi import FastAPI, Query
4
+
5
+ # Initialize FastAPI
6
+ app = FastAPI()
7
+
8
+ # Load tokenizer and model
9
+ tokenizer = GPT2Tokenizer.from_pretrained('gpt2')
10
+ model = GPT2LMHeadModel.from_pretrained('gpt2')
11
+
12
+ # Define endpoint for generating text
13
+ @app.get("/generate/")
14
+ def generate_text(prompt: str = Query(..., description="The input prompt for generating text")):
15
+ # Tokenize input text
16
+ input_ids = tokenizer.encode(prompt, return_tensors='pt')
17
+
18
+ # Generate output
19
+ output = model.generate(input_ids, max_length=100, num_return_sequences=1, pad_token_id=tokenizer.eos_token_id)
20
+
21
+ # Decode and return generated text
22
+ generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
23
+ return {"generated_text": generated_text}
24
+
25
+ # Run the FastAPI application
26
+ if __name__ == "__main__":
27
+ import uvicorn
28
+ uvicorn.run(app, host="0.0.0.0", port=8000)