Deployed successfully on SageMaker, but the output is truncated.

#2
by raafat1983 - opened

I manage to delpoy Yi-6B-Chat on Amazon SageMAker and successfully did inference , however but the output is truncated. How I can increase the output token

[# send request

predictor.predict({
"inputs": "give me a list of 10 city"
})]

[{'generated_text': 'give me a list of 10 city names in the world that start with the letter "C".Here is a list of 1'}]

Adjust the max_new_tokens parameter of GenerationConfig

lorinma changed discussion status to closed

Sign up or log in to comment