File size: 1,967 Bytes
e97665c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
# News
- [2023/09] The newest `llama2-wrapper>=0.1.14` supports llama.cpp's `gguf` models.

- [2023/08] 🔥 For developers, we offer a web server that acts as a drop-in replacement for the OpenAI API.

  - Usage: 

    ```
    python3 -m llama2_wrapper.server
    ```



- [2023/08] 🔥 For developers, we released `llama2-wrapper`  as a llama2 backend wrapper in [PYPI](https://pypi.org/project/llama2-wrapper/).

  - Install: `pip install llama2-wrapper`

  - Usage: 

    ```python
    from llama2_wrapper import LLAMA2_WRAPPER, get_prompt 
    llama2_wrapper = LLAMA2_WRAPPER(
        model_path="./models/Llama-2-7B-Chat-GGML/llama-2-7b-chat.ggmlv3.q4_0.bin",
        backend_type="llama.cpp", #options: llama.cpp, transformers, gptq
    )
    prompt = "Do you know Pytorch"
    llama2_promt = get_prompt(prompt)
    answer = llama2_wrapper(llama2_promt, temperature=0.9)
    ```

- [2023/08] 🔥 We added `benchmark.py` for users to benchmark llama2 models on their local devices.

  - Check/contribute the performance of your device in the full [performance doc](https://github.com/liltom-eth/llama2-webui/blob/main/docs/performance.md).

- [2023/07] We released **[llama2-webui](https://github.com/liltom-eth/llama2-webui)**, a gradio web UI to run Llama 2 on GPU or CPU from anywhere (Linux/Windows/Mac). 

  - Supporting models: [Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf)/[13b](https://huggingface.co/llamaste/Llama-2-13b-chat-hf)/[70b](https://huggingface.co/llamaste/Llama-2-70b-chat-hf), all [Llama-2-GPTQ](https://huggingface.co/TheBloke/Llama-2-7b-Chat-GPTQ), all [Llama-2-GGML](https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML) ...
  - Supporting model backends:  [tranformers](https://github.com/huggingface/transformers), [bitsandbytes(8-bit inference)](https://github.com/TimDettmers/bitsandbytes), [AutoGPTQ(4-bit inference)](https://github.com/PanQiWei/AutoGPTQ), [llama.cpp](https://github.com/ggerganov/llama.cpp)