Edit model card

Original model card

Buy me a coffee if you like this project ;)

Description

GGML Format model files for This project.

inference


import ctransformers

from ctransformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained(output_dir, ggml_file,
gpu_layers=32, model_type="llama")

manual_input: str = "Tell me about your last dream, please."


llm(manual_input, 
      max_new_tokens=256, 
      temperature=0.9, 
      top_p= 0.7)

Model Description

“Luna AI Llama2 Uncensored” is a Llama2 based Chat model
fine-tuned on over 40,000 long form chat discussions
This model was fine-tuned by Tap, the creator of Luna AI.
The result is an enhanced Llama2 7b model that rivals ChatGPT in performance
across a variety of tasks.

This model stands out for its long responses,
low hallucination rate, and absence of censorship mechanisms.

Model Training

The fine-tuning process was performed on an 8x a100 80GB machine.
The model was trained almost entirely on synthetic outputs.
This includes data from diverse sources which we included to create our custom dataset,
it includes multiple rounds of chats between Human & AI.

4bit GPTQ Version provided by @TheBloke - for GPU inference
GGML Version provided by @TheBloke - For CPU inference

Prompt Format

The model follows the Vicuna 1.1/ OpenChat format:

USER: I have difficulties in making friends, and I really need someone to talk to. Would you be my friend?

ASSISTANT: Of course! Friends are always here for each other. What do you like to do?

Future Plans

The model is currently being uploaded in FP16 format,
and there are plans to convert the model to GGML and GPTQ 4bit quantizations.

Benchmark Results

Task Version Metric Value Stderr
arc_challenge 0 acc_norm 0.5512 0.0146
hellaswag 0
mmlu 1 acc_norm 0.46521 0.036
truthfulqa_mc 1 mc2 0.4716 0.0155
Average - - 0.5114 0.0150

Ethical considerations

The data used to train the model is collected from various sources, mostly from the Web.
As such, it contains offensive, harmful and biased content.
We thus expect the model to exhibit such biases from the training data.

Human life

The model is not intended to inform decisions about matters central to human life,
and should not be used in such a way.

Risks and harms

Risks and harms of large language models include the generation of harmful, offensive or biased content.
These models are often prone to generating incorrect information, sometimes referred to as hallucinations.
We do not expect our model to be an exception in this regard.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Examples
Unable to determine this model's library. Check the docs .