StatPan commited on
Commit
b587822
โ€ข
1 Parent(s): a506a09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -39
README.md CHANGED
@@ -4,74 +4,61 @@ license: cc-by-nc-4.0
4
 
5
  # 42dot_LLM-PLM-1.3B_GGUF #
6
 
7
- * Model Creator: [42dot](https://huggingface.co/42dot)
8
- * original Model: [42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B)
9
 
10
- ## Description ##
11
 
12
- This repository contains the GGUF conversion and the most relevant quantizations
13
- of 42dot's
14
- [42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B) model - ready
15
- to be used with [llama.cpp](https://github.com/ggerganov/llama.cpp) and similar
16
- applications.
17
 
18
- ## Files ##
19
 
20
- In order to allow for fine-tuning (the model has the required LLaMA architecture)
21
- the original GGUF conversion has been made available
22
 
23
- * [42dot_LLM-PLM-1.3B.gguf](https://huggingface.co/rozek/42dot_LLM-SFT-1.3B_GGUF/blob/main/42dot_LLM-SFT-1.3B.gguf)
 
 
24
 
25
- From this file, the following quantizations were derived:
26
 
27
- * [42dot_LLM-SFT-1.3B-Q4_K_M](https://huggingface.co/rozek/42dot_LLM-SFT-1.3B_GGUF/blob/main/42dot_LLM-SFT-1.3B_Q5_K_M.gguf)
28
- * [42dot_LLM-SFT-1.3B-Q5_K_M](https://huggingface.co/rozek/42dot_LLM-SFT-1.3B_GGUF/blob/main/42dot_LLM-SFT-1.3B_Q5_K_M.gguf)
29
- * [42dot_LLM-SFT-1.3B-Q6_K](https://huggingface.co/rozek/42dot_LLM-SFT-1.3B_GGUF/blob/main/42dot_LLM-SFT-1.3B_Q6_0.gguf)
30
- * [42dot_LLM-SFT-1.3B-Q8_K](https://huggingface.co/rozek/42dot_LLM-SFT-1.3B_GGUF/blob/main/42dot_LLM-SFT-1.3B_Q8_0.gguf)
31
 
32
- (tell me if you need more)
33
 
34
- ## Usage Details ##
35
 
36
- Any technical details can be found on the
37
- [original model card](https://huggingface.co/42dot/42dot_LLM-SFT-1.3B)
38
- The most important ones for using this model are
39
 
40
- * context length is 4096
41
- * there does not seem to be a specific prompt structure - just provide the text
42
- you want to be completed
43
 
44
- ### Text Completion with LLaMA.cpp ###
 
 
 
45
 
46
  For simple inferencing, use a command similar to
47
 
48
  ```
49
- ./main -m 42dot_LLM-SFT-1.3B-Q8_K.gguf --temp 0 --top-k 4 --prompt "who was Joseph Weizenbaum?"
50
  ```
51
 
52
- ### Text Tokenization with LLaMA.cpp ###
53
 
54
  To get a list of tokens, use a command similar to
55
 
56
  ```
57
- ./tokenization -m 42dot_LLM-SFT-1.3B-Q8_K.gguf --prompt "who was Joseph Weizenbaum?"
58
  ```
59
 
60
- ### Embeddings Calculation with LLaMA.cpp ###
61
 
62
  Text embeddings are calculated with a command similar to
63
 
64
  ```
65
- ./embedding -m 42dot_LLM-SFT-1.3B-Q8_K.gguf --prompt "who was Joseph Weizenbaum?"
66
  ```
67
 
68
  ## License ##
69
 
70
- The original model "_is licensed under the Creative Commons
71
- Attribution-NonCommercial 4.0 (CC BY-NC 4.0)_" - for that reason, the same
72
- license was also chosen for the conversions found in this repository.
73
-
74
- So, in order to be fair and give credits to whom they belong:
75
-
76
- * the original model was created and published by [42dot](https://huggingface.co/42dot)
77
- * besides quantization, no changes were applied to the model itself
 
4
 
5
  # 42dot_LLM-PLM-1.3B_GGUF #
6
 
7
+ * ๋ชจ๋ธ ๋งŒ๋“  ์‚ฌ๋žŒ: [42dot](https://huggingface.co/42dot)
8
+ * ์›๋ณธ ๋ชจ๋ธ: [42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B)
9
 
10
+ ## ์„ค๋ช… ##
11
 
12
+ 42dot ๋ชจ๋ธ์˜ GGUF ๊ฒฝ๋Ÿ‰ํ™” ๋ชจ๋ธ์„ ๋งŒ๋“ค์–ด ๋’€์Šต๋‹ˆ๋‹ค.
 
 
 
 
13
 
14
+ [42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B)
15
 
16
+ ## ํŒŒ์ผ ##
 
17
 
18
+ ๋งํฌ์— ์—ฐ๊ฒฐ ํ•ด๋‘์—ˆ์œผ๋‹ˆ ํ•„์š”ํ•˜์‹  ๋ถ„์€ ํ•˜ํŠธ ์ฃผ๊ณ  ์ฑ™๊ฒจ ๊ฐ€์„ธ์š”.
19
+ gguf ์›๋ณธ ํŒŒ์ผ
20
+ * [42dot_LLM-PLM-1.3B.gguf](https://huggingface.co/StatPan/42dot_LLM-PLM-1.3B_GGUF/resolve/main/ggml-model-f32.gguf?download=true)
21
 
22
+ Q4, Q8 ๊ฒฝ๋Ÿ‰ํ™” ํŒŒ์ผ
23
 
24
+ * [42dot_LLM-PLM-1.3B-Q4_K_M](https://huggingface.co/StatPan/42dot_LLM-PLM-1.3B_GGUF/resolve/main/gguf-q4_k_m.gguf?download=true)
25
+ * [42dot_LLM-PLM-1.3B-Q8_K](https://huggingface.co/StatPan/42dot_LLM-PLM-1.3B_GGUF/resolve/main/gguf-q8_k_m.gguf?download=true)
 
 
26
 
27
+ ์ด์™ธ ๋ชจ๋ธ์€ ๊ทผ๋ณธ ์—†์–ด์„œ ์˜ฌ๋ฆด๊นŒ ํ•˜๋‹ค๊ฐ€ ์•ˆ ์˜ฌ๋ฆฌ๋ ค๊ณ  ํ•ฉ๋‹ˆ๋‹ค.
28
 
29
+ ## ์‚ฌ์šฉ๋ฒ• ##
30
 
 
 
 
31
 
32
+ [์›๋ณธ](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B)
 
 
33
 
34
+ ์›๋ณธ ๋งํฌ์—์„œ ์‚ฌ์šฉ ๋ฒ•์„ ํ™•์ธํ•˜์„ธ์š”.
35
+
36
+
37
+ ### Llama.cpp๋กœ ์‚ฌ์šฉ๋ฒ• ์ƒ˜ํ”Œ ###
38
 
39
  For simple inferencing, use a command similar to
40
 
41
  ```
42
+ ./main -m gguf-q4_k_m.gguf --temp 0 --top-k 4 --prompt "who was Joseph Weizenbaum?"
43
  ```
44
 
45
+ ### Llama.cpp๋กœ ํ† ํฌ๋‚˜์ด์ง• ์ƒ˜ํ”Œ ###
46
 
47
  To get a list of tokens, use a command similar to
48
 
49
  ```
50
+ ./tokenization -m gguf-q4_k_m.gguf --prompt "who was Joseph Weizenbaum?"
51
  ```
52
 
53
+ ### Llama.cpp๋กœ ์ž„๋ฒ ๋”ฉ ์ƒ˜ํ”Œ ###
54
 
55
  Text embeddings are calculated with a command similar to
56
 
57
  ```
58
+ ./embedding -m gguf-q4_k_m.gguf --prompt "who was Joseph Weizenbaum?"
59
  ```
60
 
61
  ## License ##
62
 
63
+ ์›๋ณธ ๋ชจ๋ธ ๋ผ์ด์„ผ์Šค๋Š” Creative Commons Attribution-NonCommercial 4.0 (CC BY-NC 4.0)
64
+ ์ฐธ๊ณ ํ•˜์„ธ์š”.