can't load dengcunqin/speech_paraformer-large_asr_nat-zh-cantonese-en-16k-vocab8501-online/model.pt

#3
by JminJ - opened

can't load model.pt from below code.

model = torch.load("/model/Step-Audio-Tokenizer/dengcunqin/speech_paraformer-large_asr_nat-zh-cantonese-en-16k-vocab8501-online/model.pt")

I'm setting from docker following https://github.com/stepfun-ai/Step-Audio Readme.md file.

model ckpt is saved in /model dir..
image.png

here is my container's pip list result.
Package Version


accelerate 1.3.0
aiofiles 23.2.1
aliyun-python-sdk-core 2.16.0
aliyun-python-sdk-kms 2.16.5
annotated-types 0.7.0
antlr4-python3-runtime 4.9.3
anyio 4.8.0
audioread 3.0.1
awscli 1.22.34
blinker 1.4
botocore 1.23.34
certifi 2020.6.20
cffi 1.17.1
chardet 4.0.0
charset-normalizer 3.4.1
click 8.1.8
colorama 0.4.4
coloredlogs 15.0.1
conformer 0.3.2
crcmod 1.7
cryptography 3.4.8
dbus-python 1.2.18
decorator 5.2.1
diffusers 0.32.2
distro 1.7.0
distro-info 1.1+ubuntu0.2
docutils 0.17.1
editdistance 0.8.1
einops 0.8.1
exceptiongroup 1.2.2
fastapi 0.115.11
ffmpy 0.5.0
filelock 3.17.0
flatbuffers 25.2.10
fsspec 2025.2.0
funasr 1.2.4
gradio 5.20.1
gradio_client 1.7.2
groovy 0.1.2
h11 0.14.0
httpcore 1.0.7
httplib2 0.20.2
httpx 0.28.1
huggingface-hub 0.29.2
humanfriendly 10.0
hydra-core 1.3.2
HyperPyYAML 1.2.2
idna 3.3
importlib-metadata 4.6.4
jaconv 0.4.0
jamo 0.4.1
jeepney 0.7.1
jieba 0.42.1
Jinja2 3.1.6
jmespath 0.10.0
joblib 1.4.2
kaldiio 2.18.1
keyring 23.5.0
launchpadlib 1.10.16
lazr.restfulclient 0.14.4
lazr.uri 1.0.6
lazy_loader 0.4
librosa 0.10.2.post1
llvmlite 0.44.0
markdown-it-py 3.0.0
MarkupSafe 2.1.5
mdurl 0.1.2
modelscope 1.23.2
more-itertools 8.10.0
mpmath 1.3.0
msgpack 1.1.0
networkx 3.4.2
numba 0.61.0
numpy 1.26.4
nvidia-cublas-cu12 12.1.3.1
nvidia-cuda-cupti-cu12 12.1.105
nvidia-cuda-nvrtc-cu12 12.1.105
nvidia-cuda-runtime-cu12 12.1.105
nvidia-cudnn-cu12 8.9.2.26
nvidia-cufft-cu12 11.0.2.54
nvidia-curand-cu12 10.3.2.106
nvidia-cusolver-cu12 11.4.5.107
nvidia-cusparse-cu12 12.1.0.106
nvidia-nccl-cu12 2.20.5
nvidia-nvjitlink-cu12 12.8.93
nvidia-nvtx-cu12 12.1.105
oauthlib 3.2.0
olefile 0.46
omegaconf 2.3.0
onnxruntime-gpu 1.17.0
openai-whisper 20231117
orjson 3.10.15
oss2 2.19.1
packaging 24.2
pandas 2.2.3
pillow 11.1.0
pip 25.0.1
platformdirs 4.3.6
pooch 1.8.2
protobuf 5.29.3
psutil 7.0.0
pyasn1 0.4.8
pycparser 2.22
pycryptodome 3.21.0
pydantic 2.10.6
pydantic_core 2.27.2
pydub 0.25.1
Pygments 2.19.1
PyGObject 3.42.1
PyJWT 2.3.0
pynndescent 0.5.13
pyparsing 2.4.7
python-apt 2.4.0+ubuntu4
python-dateutil 2.9.0.post0
python-multipart 0.0.20
pytorch-wpe 0.0.1
pytz 2025.1
PyYAML 5.4.1
regex 2024.11.6
requests 2.32.3
rich 13.9.4
roman 3.3
rsa 4.8
ruamel.yaml 0.18.10
ruamel.yaml.clib 0.2.12
ruff 0.9.9
s3transfer 0.5.0
safehttpx 0.1.6
safetensors 0.5.3
scikit-learn 1.6.1
scipy 1.15.2
SecretStorage 3.3.1
semantic-version 2.10.0
sentencepiece 0.2.0
setuptools 75.8.2
shellingham 1.5.4
six 1.16.0
sniffio 1.3.1
soundfile 0.13.1
sox 1.5.0
soxr 0.5.0.post1
ssh-import-id 5.11
starlette 0.46.0
sympy 1.13.3
tensorboardX 2.6.2.2
threadpoolctl 3.5.0
tiktoken 0.9.0
tokenizers 0.21.0
tomlkit 0.13.2
torch 2.3.1
torch-complex 0.4.4
torchaudio 2.3.1
torchvision 0.18.1
tqdm 4.67.1
transformers 4.48.3
triton 2.3.1
typer 0.15.2
typing_extensions 4.12.2
tzdata 2025.1
umap-learn 0.5.7
unattended-upgrades 0.1
urllib3 1.26.5
uvicorn 0.34.0
wadllib 1.3.6
websockets 15.0.1
wheel 0.45.1
zipp 1.0.0

JminJ changed discussion status to closed

Sign up or log in to comment