Sagemaker Payload limit issue (413)

#24
by MLLife - opened

please refer to the issue detailed here; https://discuss.huggingface.co/t/deploying-open-ais-whisper-on-sagemaker/24761/54?u=mllife

basically, the currently sagemaker have a max payload set to 5 MB, and there is no way around it on how the current code for whisper is streaming the file to the end-point using just audio_path as input; which makes this model nearly useless for sagemaker deployment.

if someone has done custom inference.py which loads file from s3_path at the endpoint itself and later processes it, please share

Sign up or log in to comment