updated README.md on how to make changes

#17
by absy - opened
Files changed (1) hide show
  1. README.md +16 -0
README.md CHANGED
@@ -123,6 +123,22 @@ If you want to have a chat-style conversation, replace the `-p <PROMPT>` argumen
123
 
124
  Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
125
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
126
  <!-- footer start -->
127
  ## Discord
128
 
 
123
 
124
  Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
125
 
126
+ ## How to make changes.
127
+
128
+ Downloading GGML files from repositories can be done in a more efficient way than using git clone. Here's how:
129
+
130
+ 1. Avoid Git Clone: It's not recommended to use git clone for GGML repositories as it downloads all files, resulting in unnecessary data storage.
131
+ 2. Use wget or Download Client: Instead, use the wget command or any download client to directly grab the specific file you need. This prevents excessive data usage and storage.
132
+ 3. Python Code Option: If you prefer Python, you can use the hf_hub_download function from huggingface_hub. Here's an example:
133
+
134
+ ```
135
+ from huggingface_hub import hf_hub_download
136
+ hf_hub_download(repo_id='TheBloke/Llama-2-7B-Chat-GGML', filename='llama-2-7b-chat.ggmlv3.q4_K_M.bin')
137
+ #This code snippet downloads the specified GGML file directly.
138
+ ```
139
+
140
+ 4. Inference from Python: If you plan to perform inference from Python, the ctransformers Python library is a convenient option. It handles automatic file downloads for GGML inference.
141
+
142
  <!-- footer start -->
143
  ## Discord
144