Edit model card

Phi 3 Model with Extended Vocabulary and Fine-Tuning for Japanese

Overview

This project is a proof of concept that extends the base vocabulary of the Phi 3 model and then applies supervised fine-tuning to teach it a new language (Japanese). Despite using a very small custom dataset, the improvement in Japanese language understanding is substantial.

Model Details

  • Base Model: Phi 3
  • Objective: Extend the base vocabulary and fine-tune for Japanese language understanding.
  • Dataset: Custom dataset of 1,000 entries generated using ChatGPT-4.
  • Language: Japanese

Dataset

The dataset used for this project was generated with the assistance of ChatGPT-4. It comprises 1,000 entries, carefully curated to cover a diverse range of topics and linguistic structures.

Training

Vocabulary Extension

The base vocabulary of the Phi 3 model was extended to include new Japanese tokens. This was a crucial step to enable the model to comprehend and generate Japanese text more effectively.

Fine-Tuning

Supervised fine-tuning was performed on the extended model using the custom dataset. Despite the small dataset size, the model showed significant improvement in understanding and generating Japanese text.

Results

Even with the limited dataset and vocabulary size, the fine-tuned model demonstrated substantial improvements over the base model in terms of Japanese language understanding and generation.

Future Work

  1. Dataset Expansion: Increase the size and diversity of the dataset to further enhance model performance.
  2. Evaluation: Conduct comprehensive evaluation and benchmarking against standard Japanese language tasks.
  3. Optimization: Optimize the model for better performance and efficiency.
Downloads last month

-

Downloads are not tracked for this model. How to track
Unable to determine this model's library. Check the docs .