dahara1 commited on
Commit
3d5dc8c
1 Parent(s): 32beddb

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -0
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ inference: false
3
+ language:
4
+ - ja
5
+ - en
6
+ ---
7
+ # webbigdata/ALMA-7B-Ja-GPTQ-Ja-En
8
+
9
+ Original ALMA Model [ALMA-7B](https://huggingface.co/haoranxu/ALMA-7B). (26.95GB) is a new paradigm translation model.
10
+
11
+ [ALMA-7B-Ja-V2-GPTQ-Ja-En](https://huggingface.co/webbigdata/ALMA-7B-Ja-V2-GPTQ-Ja-En) is a machine translation model that uses ALMA's learning method to translate Japanese to English.(13.3GB)
12
+
13
+ This model is GPTQ quantized version model that reduces model size(3.9GB) and memory usage, although the performance is probably lower.
14
+ And translation ability for languages other than Japanese and English has deteriorated significantly.
15
+
16
+ [Free Colab Sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_V2_GPTQ_Ja_En_Free_Colab_sample.ipynb)
17
+
18
+ If you want to translate the entire file at once, try Colab below.
19
+ [ALMA_7B_Ja_V2_GPTQ_Ja_En_batch_translation_sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_V2_GPTQ_Ja_En_batch_translation_sample.ipynb)
20
+
21
+ if you enconter error below.
22
+ ```RuntimeError: probability tensor contains either `inf`, `nan` or element < 0```
23
+ It's mean your memory is not enough. decrease your num_beams or token size.
24
+
25
+
26
+ **ALMA** (**A**dvanced **L**anguage **M**odel-based tr**A**nslator) is an LLM-based translation model, which adopts a new translation model paradigm: it begins with fine-tuning on monolingual data and is further optimized using high-quality parallel data. This two-step fine-tuning process ensures strong translation performance.
27
+ Please find more details in their [paper](https://arxiv.org/abs/2309.11674).
28
+ ```
29
+ @misc{xu2023paradigm,
30
+ title={A Paradigm Shift in Machine Translation: Boosting Translation Performance of Large Language Models},
31
+ author={Haoran Xu and Young Jin Kim and Amr Sharaf and Hany Hassan Awadalla},
32
+ year={2023},
33
+ eprint={2309.11674},
34
+ archivePrefix={arXiv},
35
+ primaryClass={cs.CL}
36
+ }
37
+ ```
38
+
39
+ ## about this work
40
+ - **This work was done by :** [webbigdata](https://webbigdata.jp/).