Text Generation
Transformers
Safetensors
Indonesian
bloom
conversational
Inference Endpoints
text-generation-inference
haidlir commited on
Commit
61ea9c1
1 Parent(s): 376034c

Update README.md (#2)

Browse files

- Update README.md (6db5b387dc0536e8ffdb6e805971de02049dd79c)

Files changed (1) hide show
  1. README.md +14 -6
README.md CHANGED
@@ -20,11 +20,19 @@ pipeline_tag: text-generation
20
  - https://huggingface.co/datasets/jakartaresearch/indoqa
21
 
22
 
23
- **Task**: Chat or Conversational
24
- **Input**: User's prompt containing chat templated text in string format
25
- **Output**: Model's generated text in string format
 
 
 
 
 
26
 
27
  **Experiment**:
28
- - Use bos and eos token to replace <|im_start|> and <|im_end|> in ChatML. (Inspired by: https://asmirnov.xyz/doppelganger)
29
- - Penggunaan padding dan truncation sesuai max_length.
30
- - Max length = 256, karena telah mengkonsumsi 33.7 GB.
 
 
 
 
20
  - https://huggingface.co/datasets/jakartaresearch/indoqa
21
 
22
 
23
+ **Task**:
24
+ Chat or Conversational
25
+
26
+ **Input**:
27
+ User's prompt containing chat templated text in string format
28
+
29
+ **Output**:
30
+ Model's generated text in string format
31
 
32
  **Experiment**:
33
+ - Use bos_token and eos_token to replace <|im_start|> and <|im_end|> in ChatML. (Inspired by: https://asmirnov.xyz/doppelganger)
34
+ - Use left padding and left truncation to conform to max_length.
35
+ - Set max_length = 256 in the training process, which consumes 33.7 GB of memory.
36
+
37
+ **Notebook**:
38
+ - https://drive.google.com/file/d/11FiaWxGt2HxUirZrHTNLaVmiqrUwejwV/view?usp=drive_link