Techbite commited on
Commit
3c4eeeb
·
1 Parent(s): f730649
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -108,4 +108,4 @@ This implementation includes several optimizations for systems with 16-19GB RAM
108
  - The embedding and retrieval components work efficiently even on limited hardware
109
  - Response generation speed depends on the model size and available GPU memory
110
  - For optimal performance with 8GB GPU, stick with Phi-2 model
111
- - For faster responses with less accuracy, use TinyLlama-1.1B
 
108
  - The embedding and retrieval components work efficiently even on limited hardware
109
  - Response generation speed depends on the model size and available GPU memory
110
  - For optimal performance with 8GB GPU, stick with Phi-2 model
111
+ - For faster responses with less accuracy, use TinyLlama-1.1B -->