P0x0 commited on
Commit
b65d69d
1 Parent(s): 3cde4d6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -13
README.md CHANGED
@@ -38,16 +38,5 @@ Astra-v1-12B can be used directly for a wide range of NLP tasks, including:
38
  ### Out-of-Scope Use
39
 
40
  Astra-v1-12B is not intended for real-time decision-making in critical applications or generating harmful or biased content.
41
- ## How to Get Started with the Model
42
-
43
- ```python
44
- from transformers import AutoModelForCausalLM, AutoTokenizer
45
-
46
- tokenizer = AutoTokenizer.from_pretrained("P0x0/astra-v1-12b")
47
- model = AutoModelForCausalLM.from_pretrained("P0x0/astra-v1-12b")
48
-
49
- input_text = "Explain the theory of relativity in simple terms."
50
- inputs = tokenizer(input_text, return_tensors="pt")
51
- outputs = model.generate(**inputs)
52
-
53
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
 
38
  ### Out-of-Scope Use
39
 
40
  Astra-v1-12B is not intended for real-time decision-making in critical applications or generating harmful or biased content.
41
+ ## How to Get Started with the quantized model
42
+ To run the quantized version of the model, you can use [KoboldCPP](https://github.com/LostRuins/koboldcpp), which allows you to run quantized GGUF models locally.