File size: 1,115 Bytes
30d9278
 
e80a3c3
30d9278
 
 
e80a3c3
30d9278
e80a3c3
30d9278
75a22b4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31

# Bio-SIEVE
[Arxiv](https://arxiv.org/abs/2308.06610) [Github](https://github.com/GateNLP/Bio-SIEVE)

The single-task Bio-SIEVE optimised for Inclusion/Exclusion classification of biomedical literature given the systematic review's objectives and selection criteria.

These are LoRA weights, having continued training from Guanaco7B. The base LLaMA model will be automatically downloaded when used with the generate_cli.py script from our [Github](https://github.com/GateNLP/Bio-SIEVE).

Also see the [multi-task variant](https://huggingface.co/GateNLP/Bio-SIEVE-Multi) for exclusion reasoning and PIO extraction. Variants trained without instruction tuned pretraining TBD.

---
library_name: peft
---
## Training procedure


The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: float32
### Framework versions


- PEFT 0.4.0.dev0