prithivMLmods commited on
Commit
49e382e
·
verified ·
1 Parent(s): c979521

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -12
README.md CHANGED
@@ -34,11 +34,14 @@ With its robust natural language processing capabilities, **Codepy 3B Deep Think
34
  | `tokenizer.json` | 17.2 MB | Full tokenizer vocabulary and merges. | Uploaded (LFS) |
35
  | `tokenizer_config.json` | 57.5 kB | Tokenizer configuration details. | Uploaded |
36
 
37
- | Model Type | Size | Context Length | Link |
38
- |------------|------|----------------|------|
39
- | GGUF | 3B | - | [🤗 Codepy-Deepthink-3B-GGUF](https://huggingface.co/prithivMLmods/Codepy-Deepthink-3B-GGUF) |
 
 
 
 
40
 
41
- ---
42
  # **<span style="color:red;">Sample Deepthink Inference</span>**
43
 
44
  >>> Develop a Python program to generate random passwords that consist of 8 characters. Not applicable
@@ -143,14 +146,6 @@ To customize the password length, modify the `password_length` variable in the s
143
  - This implementation uses Python's `random` module, which is suitable for general-purpose randomness. For cryptographically secure passwords, consider using the `secrets` module.
144
  - The character set includes spaces for additional complexity, but you can modify the `characters` string to include other symbols (e.g., `!@#$%^&*`).
145
 
146
- ---
147
- # **Run with LM Studio**
148
-
149
- | Feature Run | Details |
150
- |--------------------------|-----------------------------------------------------------------------------------------------|
151
- | **Run with LM Studio** | https://lmstudio.ai/ |
152
- | **Demo on LM Studio** | https://drive.google.com/file/d/1CHdfjYrwMnk9ACvS40Abfy3xNXnCubKG/view?usp=sharing |
153
-
154
  # **Model Architecture**
155
 
156
  Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
 
34
  | `tokenizer.json` | 17.2 MB | Full tokenizer vocabulary and merges. | Uploaded (LFS) |
35
  | `tokenizer_config.json` | 57.5 kB | Tokenizer configuration details. | Uploaded |
36
 
37
+ # **Run with LM Studio**
38
+
39
+ | Feature Run | Details |
40
+ |--------------------------|-----------------------------------------------------------------------------------------------|
41
+ | **Run with LM Studio** | https://lmstudio.ai/ |
42
+ | **Demo on LM Studio** | https://drive.google.com/file/d/1CHdfjYrwMnk9ACvS40Abfy3xNXnCubKG/view?usp=sharing |
43
+ | **Codepy-Deepthink-3B-GGUF** | https://huggingface.co/prithivMLmods/Codepy-Deepthink-3B-GGUF |
44
 
 
45
  # **<span style="color:red;">Sample Deepthink Inference</span>**
46
 
47
  >>> Develop a Python program to generate random passwords that consist of 8 characters. Not applicable
 
146
  - This implementation uses Python's `random` module, which is suitable for general-purpose randomness. For cryptographically secure passwords, consider using the `secrets` module.
147
  - The character set includes spaces for additional complexity, but you can modify the `characters` string to include other symbols (e.g., `!@#$%^&*`).
148
 
 
 
 
 
 
 
 
 
149
  # **Model Architecture**
150
 
151
  Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.