kazssym's picture
Update README
45ae79e
|
raw
history blame
1.02 kB
---
license: cc-by-sa-4.0
tags:
- causal-lm
---
# Description
This is a test model for my [test script](https://github.com/kazssym/stablelm-study-2).
It was exported from [stabilityai/stablelm-3b-4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t) to ONNX with a [modified version](https://github.com/huggingface/optimum/pull/1719) of Hugging Face Optimum.
It is quite possible to have problems.
This model does not include a tokenizer.
# Export command
This model was exported with the following command:
```
optimum-cli export onnx --model stabilityai/stablelm-3b-4e1t --trust-remote-code --device cpu --optimize O1 output/onnx-fp32/
```
It requires [Transformers](https://github.com/huggingface/transformers) 4.38 or later to export.
# Output from Optimum CLI
```
The ONNX export succeeded with the warning: The maximum absolute difference between the output of the reference model and the ONNX exported model is not within the set tolerance 1e-05:
- logits: max diff = 2.6553869247436523e-05.
```