File size: 1,801 Bytes
e65ae1d
 
056dedb
 
 
 
 
eb8b6a3
 
e65ae1d
 
 
056dedb
e65ae1d
 
 
 
 
 
 
 
 
056dedb
e65ae1d
056dedb
 
 
 
e65ae1d
 
 
 
 
056dedb
e65ae1d
 
 
056dedb
e65ae1d
 
 
 
056dedb
e65ae1d
 
 
 
 
 
056dedb
 
e65ae1d
056dedb
e65ae1d
056dedb
 
 
e65ae1d
056dedb
 
e65ae1d
056dedb
e65ae1d
056dedb
e65ae1d
056dedb
e65ae1d
056dedb
e65ae1d
eb8b6a3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
library_name: transformers
datasets:
- ahmed000000000/cybersec
- dzakwan/cybersec
language:
- en
tags:
- conversational
---
# Model Card for Model ID

Works as a cyber assistant.



## Model Details

### Model Description

<!-- Provide a longer summary of what this model is. -->

This is the model card of a 🤗 transformers model that has been pushed on the Hub.

- **Developed by:**  <a href="https://github.com/Zardian18">Zardian18</a>
- **Model type:** GPT2
- **Language(s) (NLP):** English
- **Finetuned from model [optional]:** <a href="https://huggingface.co/openai-community/gpt2-medium">OpenAi GPT2-medium</a>

### Model Sources [optional]

<!-- Provide the basic links for the model. -->

- **Repository:** <a href="https://github.com/Zardian18/CyberAssist">Github repo</a>

## Uses

Can be used to handle and solve basic cybersec queries.


## Bias, Risks, and Limitations

Currently it is fine-tuned on GPT2, which is good but not comparable to state of the art LLMs and Transformers. Moreover, the dataset is small.

[More Information Needed]


## How to Get Started with the Model

```# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="Zardian/Cyber_assist3.0")

```
```# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("Zardian/Cyber_assist3.0")
model = AutoModelForCausalLM.from_pretrained("Zardian/Cyber_assist3.0")

```

## Training Details

### Training Data

<a href="https://huggingface.co/datasets/ahmed000000000/cybersec">Cybersec queries and responses dataset</a> consisting of 12k enteries.

<a href="https://huggingface.co/datasets/dzakwan/cybersec">Cybersec dataset with instructions and output</a> consisting of 14k enteries.