File size: 1,801 Bytes
e65ae1d 056dedb eb8b6a3 e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d 056dedb e65ae1d eb8b6a3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 |
---
library_name: transformers
datasets:
- ahmed000000000/cybersec
- dzakwan/cybersec
language:
- en
tags:
- conversational
---
# Model Card for Model ID
Works as a cyber assistant.
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub.
- **Developed by:** <a href="https://github.com/Zardian18">Zardian18</a>
- **Model type:** GPT2
- **Language(s) (NLP):** English
- **Finetuned from model [optional]:** <a href="https://huggingface.co/openai-community/gpt2-medium">OpenAi GPT2-medium</a>
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** <a href="https://github.com/Zardian18/CyberAssist">Github repo</a>
## Uses
Can be used to handle and solve basic cybersec queries.
## Bias, Risks, and Limitations
Currently it is fine-tuned on GPT2, which is good but not comparable to state of the art LLMs and Transformers. Moreover, the dataset is small.
[More Information Needed]
## How to Get Started with the Model
```# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-generation", model="Zardian/Cyber_assist3.0")
```
```# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Zardian/Cyber_assist3.0")
model = AutoModelForCausalLM.from_pretrained("Zardian/Cyber_assist3.0")
```
## Training Details
### Training Data
<a href="https://huggingface.co/datasets/ahmed000000000/cybersec">Cybersec queries and responses dataset</a> consisting of 12k enteries.
<a href="https://huggingface.co/datasets/dzakwan/cybersec">Cybersec dataset with instructions and output</a> consisting of 14k enteries. |