Is your params can run with default APK ?
Hi Kmadsen,
I checked my params is totally same with yours, but can't run with default APK, can you run with it ?
It will have no response after input some words and press send button.
I renamed the params and pushed to this dir
/storage/emulated/0/Android/data/ai.mlc.mlcchat/files/vicuna-v1-7b-q4f16_0 # ls -l
total 3707480
-rw-rw---- 1 root ext_data_rw 395 2022-02-23 12:09 mlc-chat-config.json
-rw-rw---- 1 root ext_data_rw 200823 2022-02-23 12:09 ndarray-cache.json
-rw-rw---- 1 root ext_data_rw 65536000 2022-02-23 12:09 params_shard_0.bin
-rw-rw---- 1 root ext_data_rw 27066368 2022-02-23 12:09 params_shard_1.bin
...
Thanks,
Qing
Hi Qing
I made a edit to the MLCChat application so that our fork will accept unaltered model directories from huggingface. This is what the configuration looks like, notice the added model_url_directory
.
{
"model_libs": [
"RedPajama-INCITE-Chat-3B-v1-q4f16_0",
"vicuna-7b-v1.1-q4f16_0"
],
"model_list": [
{
"model_url": "https://huggingface.co/mlc-ai/mlc-chat-RedPajama-INCITE-Chat-3B-v1-q4f16_0/",
"local_id": "RedPajama-INCITE-Chat-3B-v1-q4f16_0"
},
{
"model_url": "https://huggingface.co/kmadsen/vicuna-7b-v1.1-q4f16_0/",
"model_url_directory": "params/",
"local_id": "vicuna-7b-v1.1-q4f16_0"
}
],
"add_model_samples": [
]
}
Ok, thanks for your comment. But my issues not related with name, I directly change name from vicuna-7b-v1.1-q4f16_0 to vicuna-v1-7b-q4f16_0 is also ok.
I checked ndarray-cache.json with default ndarray-cache.json seems a little different, my ndarray-cache.json first param as below:
"dataPath": "params_shard_0.bin",
"format": "raw-shard",
"nbytes": 65536000,
"records": [
{
"name": "param_0",
But default is :
"dataPath": "params_shard_0.bin",
"format": "raw-shard",
"nbytes": 65536000,
"records": [
{
"name": "param_67",
Default can work, but my generated params not work, will check further.
Thanks again for your comment.
Thanks,
Qing
The reason I uploaded this was to keep track of vicuna-7b-v1.1-q4f16_0-android.tar
. For some reason it's not uploaded to https://github.com/mlc-ai/binary-mlc-llm-libs
I was facing similar issues as you except it was when using the pre-builds in https://huggingface.co/mlc-ai. This upload allows us to build the android app and load the params from a single place.
Am considering this model all temporary, but if you're looking to customize an android app for testing; then this model will allow you to skip the custom TVM+OpenCL. But you'll still need to prepare_libs.sh
.