lemon-mint's picture
Update README.md
e6258f3 verified
|
raw
history blame
1.54 kB
metadata
language:
  - ko
  - en
license: gemma
library_name: transformers
tags:
  - korean
  - gemma
  - pytorch
pipeline_tag: text-generation
base_model: google/gemma-1.1-7b-it

image/jpeg

Gemma Ko 7B Instruct v0.71

  • Eval Loss: 1.51977
  • Train Loss: 0.48541
  • lr: 5e-5
  • optimizer: adamw
  • lr_scheduler_type: cosine

Model Details

Model Description

The Gemma Ko 7B Instruct v0.71 model is designed for generating human-like text in the Korean language. It can be used for a variety of natural language processing tasks, such as language translation, text summarization, question answering, and conversation generation. This model is particularly well-suited for applications that require high-quality, coherent, and contextually relevant Korean text generation.

Limitations and Ethical Considerations

As Gemma Ko 7B has been trained on extensive web data, biases present in the training data may be reflected in the model. Additionally, there is a possibility that it may generate sentences containing errors or incorrect information. Therefore, rather than blindly trusting the model's output, it is necessary to refer to it with caution.