shantipriya commited on
Commit
3d13e2f
1 Parent(s): e4f9382

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -26,6 +26,21 @@ For more details about the model, data, training procedure, and evaluations, go
26
  Sample inference script.
27
 
28
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  from unsloth import FastLanguageModel
30
  import torch
31
  max_seq_length = 2048
@@ -54,7 +69,7 @@ FastLanguageModel.for_inference(model)
54
  inputs = tokenizer(
55
  [
56
  alpaca_prompt.format(
57
- "ଓଡିଶାର ରାଜଧାନୀ କ’ଣ?", # instruction
58
  "", # input
59
  "", # output - leave this blank for generation!
60
  )
 
26
  Sample inference script.
27
 
28
  ```
29
+ #Install Unsloth
30
+ %%capture
31
+ import torch
32
+ major_version, minor_version = torch.cuda.get_device_capability()
33
+ # Must install separately since Colab has torch 2.2.1, which breaks packages
34
+ !pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
35
+ if major_version >= 8:
36
+ # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)
37
+ !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes
38
+ else:
39
+ # Use this for older GPUs (V100, Tesla T4, RTX 20xx)
40
+ !pip install --no-deps xformers trl peft accelerate bitsandbytes
41
+ pass
42
+
43
+
44
  from unsloth import FastLanguageModel
45
  import torch
46
  max_seq_length = 2048
 
69
  inputs = tokenizer(
70
  [
71
  alpaca_prompt.format(
72
+ "କୋଭିଡ୍ 19 ର ଲକ୍ଷଣଗୁଡ଼ିକ କ’ଣ?", # instruction
73
  "", # input
74
  "", # output - leave this blank for generation!
75
  )