Promptist Inference

#1
by mmilunovic - opened
Microsoft org

After reading the paper couple times, I'm still not clear on what actually happens during inference?
I understand we call GPT-2 to rephrase the original prompt, but what happens after that?
What's the performance of the entire inference pipeline (inference time for example) and is there a way to boost it?

Sign up or log in to comment