LoWRA-Bench / README.md
Eliahu's picture
Upload dataset
9a5c5a4 verified
|
raw
history blame
No virus
16.9 kB
metadata
dataset_info:
  - config_name: mistral-7b-v0.1-dpo
    features:
      - name: task_name
        dtype: string
      - name: layer_model
        dtype: string
      - name: layer_name
        dtype: string
      - name: pre_ft_name
        dtype: string
      - name: pre_ft_weight
        sequence:
          sequence: float32
      - name: lora_0_name
        dtype: string
      - name: lora_0_A_weight
        sequence:
          sequence: float32
      - name: lora_0_B_weight
        sequence:
          sequence: float32
      - name: lora_0_rank
        dtype: int64
      - name: lora_0_alpha
        dtype: int64
      - name: lora_1_name
        dtype: string
      - name: lora_1_A_weight
        sequence:
          sequence: float32
      - name: lora_1_B_weight
        sequence:
          sequence: float32
      - name: lora_1_rank
        dtype: int64
      - name: lora_1_alpha
        dtype: int64
      - name: lora_2_name
        dtype: string
      - name: lora_2_A_weight
        sequence:
          sequence: float32
      - name: lora_2_B_weight
        sequence:
          sequence: float32
      - name: lora_2_rank
        dtype: int64
      - name: lora_2_alpha
        dtype: int64
      - name: lora_3_name
        dtype: string
      - name: lora_3_A_weight
        sequence:
          sequence: float32
      - name: lora_3_B_weight
        sequence:
          sequence: float32
      - name: lora_3_rank
        dtype: int64
      - name: lora_3_alpha
        dtype: int64
      - name: lora_4_name
        dtype: string
      - name: lora_4_A_weight
        sequence:
          sequence: float32
      - name: lora_4_B_weight
        sequence:
          sequence: float32
      - name: lora_4_rank
        dtype: int64
      - name: lora_4_alpha
        dtype: int64
      - name: lora_5_name
        dtype: string
      - name: lora_5_A_weight
        sequence:
          sequence: float32
      - name: lora_5_B_weight
        sequence:
          sequence: float32
      - name: lora_5_rank
        dtype: int64
      - name: lora_5_alpha
        dtype: int64
      - name: lora_6_name
        dtype: string
      - name: lora_6_A_weight
        sequence:
          sequence: float32
      - name: lora_6_B_weight
        sequence:
          sequence: float32
      - name: lora_6_rank
        dtype: int64
      - name: lora_6_alpha
        dtype: int64
      - name: lora_7_name
        dtype: string
      - name: lora_7_A_weight
        sequence:
          sequence: float32
      - name: lora_7_B_weight
        sequence:
          sequence: float32
      - name: lora_7_rank
        dtype: int64
      - name: lora_7_alpha
        dtype: int64
      - name: lora_8_name
        dtype: string
      - name: lora_8_A_weight
        sequence:
          sequence: float32
      - name: lora_8_B_weight
        sequence:
          sequence: float32
      - name: lora_8_rank
        dtype: int64
      - name: lora_8_alpha
        dtype: int64
      - name: lora_9_name
        dtype: string
      - name: lora_9_A_weight
        sequence:
          sequence: float32
      - name: lora_9_B_weight
        sequence:
          sequence: float32
      - name: lora_9_rank
        dtype: int64
      - name: lora_9_alpha
        dtype: int64
      - name: lora_10_name
        dtype: string
      - name: lora_10_A_weight
        sequence:
          sequence: float32
      - name: lora_10_B_weight
        sequence:
          sequence: float32
      - name: lora_10_rank
        dtype: int64
      - name: lora_10_alpha
        dtype: int64
      - name: lora_11_name
        dtype: string
      - name: lora_11_A_weight
        sequence:
          sequence: float32
      - name: lora_11_B_weight
        sequence:
          sequence: float32
      - name: lora_11_rank
        dtype: int64
      - name: lora_11_alpha
        dtype: int64
      - name: lora_12_name
        dtype: string
      - name: lora_12_A_weight
        sequence:
          sequence: float32
      - name: lora_12_B_weight
        sequence:
          sequence: float32
      - name: lora_12_rank
        dtype: int64
      - name: lora_12_alpha
        dtype: int64
      - name: lora_13_name
        dtype: string
      - name: lora_13_A_weight
        sequence:
          sequence: float32
      - name: lora_13_B_weight
        sequence:
          sequence: float32
      - name: lora_13_rank
        dtype: int64
      - name: lora_13_alpha
        dtype: int64
      - name: lora_14_name
        dtype: string
      - name: lora_14_A_weight
        sequence:
          sequence: float32
      - name: lora_14_B_weight
        sequence:
          sequence: float32
      - name: lora_14_rank
        dtype: int64
      - name: lora_14_alpha
        dtype: int64
    splits:
      - name: train
        num_bytes: 8661875544
        num_examples: 128
    download_size: 3419054382
    dataset_size: 8661875544
  - config_name: mistral-7b-v0.1-sft
    features:
      - name: task_name
        dtype: string
      - name: layer_model
        dtype: string
      - name: layer_name
        dtype: string
      - name: pre_ft_name
        dtype: string
      - name: pre_ft_weight
        sequence:
          sequence: float32
      - name: lora_0_name
        dtype: string
      - name: lora_0_A_weight
        sequence:
          sequence: float32
      - name: lora_0_B_weight
        sequence:
          sequence: float32
      - name: lora_0_rank
        dtype: int64
      - name: lora_0_alpha
        dtype: int64
      - name: lora_1_name
        dtype: string
      - name: lora_1_A_weight
        sequence:
          sequence: float32
      - name: lora_1_B_weight
        sequence:
          sequence: float32
      - name: lora_1_rank
        dtype: int64
      - name: lora_1_alpha
        dtype: int64
      - name: lora_2_name
        dtype: string
      - name: lora_2_A_weight
        sequence:
          sequence: float32
      - name: lora_2_B_weight
        sequence:
          sequence: float32
      - name: lora_2_rank
        dtype: int64
      - name: lora_2_alpha
        dtype: int64
      - name: lora_3_name
        dtype: string
      - name: lora_3_A_weight
        sequence:
          sequence: float32
      - name: lora_3_B_weight
        sequence:
          sequence: float32
      - name: lora_3_rank
        dtype: int64
      - name: lora_3_alpha
        dtype: int64
      - name: lora_4_name
        dtype: string
      - name: lora_4_A_weight
        sequence:
          sequence: float32
      - name: lora_4_B_weight
        sequence:
          sequence: float32
      - name: lora_4_rank
        dtype: int64
      - name: lora_4_alpha
        dtype: int64
      - name: lora_5_name
        dtype: string
      - name: lora_5_A_weight
        sequence:
          sequence: float32
      - name: lora_5_B_weight
        sequence:
          sequence: float32
      - name: lora_5_rank
        dtype: int64
      - name: lora_5_alpha
        dtype: int64
      - name: lora_6_name
        dtype: string
      - name: lora_6_A_weight
        sequence:
          sequence: float32
      - name: lora_6_B_weight
        sequence:
          sequence: float32
      - name: lora_6_rank
        dtype: int64
      - name: lora_6_alpha
        dtype: int64
      - name: lora_7_name
        dtype: string
      - name: lora_7_A_weight
        sequence:
          sequence: float32
      - name: lora_7_B_weight
        sequence:
          sequence: float32
      - name: lora_7_rank
        dtype: int64
      - name: lora_7_alpha
        dtype: int64
      - name: lora_8_name
        dtype: string
      - name: lora_8_A_weight
        sequence:
          sequence: float32
      - name: lora_8_B_weight
        sequence:
          sequence: float32
      - name: lora_8_rank
        dtype: int64
      - name: lora_8_alpha
        dtype: int64
      - name: lora_9_name
        dtype: string
      - name: lora_9_A_weight
        sequence:
          sequence: float32
      - name: lora_9_B_weight
        sequence:
          sequence: float32
      - name: lora_9_rank
        dtype: int64
      - name: lora_9_alpha
        dtype: int64
      - name: lora_10_name
        dtype: string
      - name: lora_10_A_weight
        sequence:
          sequence: float32
      - name: lora_10_B_weight
        sequence:
          sequence: float32
      - name: lora_10_rank
        dtype: int64
      - name: lora_10_alpha
        dtype: int64
      - name: lora_11_name
        dtype: string
      - name: lora_11_A_weight
        sequence:
          sequence: float32
      - name: lora_11_B_weight
        sequence:
          sequence: float32
      - name: lora_11_rank
        dtype: int64
      - name: lora_11_alpha
        dtype: int64
      - name: lora_12_name
        dtype: string
      - name: lora_12_A_weight
        sequence:
          sequence: float32
      - name: lora_12_B_weight
        sequence:
          sequence: float32
      - name: lora_12_rank
        dtype: int64
      - name: lora_12_alpha
        dtype: int64
      - name: lora_13_name
        dtype: string
      - name: lora_13_A_weight
        sequence:
          sequence: float32
      - name: lora_13_B_weight
        sequence:
          sequence: float32
      - name: lora_13_rank
        dtype: int64
      - name: lora_13_alpha
        dtype: int64
      - name: lora_14_name
        dtype: string
      - name: lora_14_A_weight
        sequence:
          sequence: float32
      - name: lora_14_B_weight
        sequence:
          sequence: float32
      - name: lora_14_rank
        dtype: int64
      - name: lora_14_alpha
        dtype: int64
    splits:
      - name: train
        num_bytes: 8661875544
        num_examples: 128
    download_size: 5791365905
    dataset_size: 8661875544
  - config_name: stable-diffusion-1.5
    features:
      - name: task_name
        dtype: string
      - name: layer_model
        dtype: string
      - name: layer_name
        dtype: string
      - name: pre_ft_name
        dtype: string
      - name: pre_ft_weight
        sequence:
          sequence: float32
      - name: lora_0_name
        dtype: string
      - name: lora_0_A_weight
        sequence:
          sequence: float32
      - name: lora_0_B_weight
        sequence:
          sequence: float32
      - name: lora_0_rank
        dtype: int64
      - name: lora_0_alpha
        dtype: float64
      - name: lora_1_name
        dtype: string
      - name: lora_1_A_weight
        sequence:
          sequence: float32
      - name: lora_1_B_weight
        sequence:
          sequence: float32
      - name: lora_1_rank
        dtype: int64
      - name: lora_1_alpha
        dtype: float64
      - name: lora_2_name
        dtype: string
      - name: lora_2_A_weight
        sequence:
          sequence: float32
      - name: lora_2_B_weight
        sequence:
          sequence: float32
      - name: lora_2_rank
        dtype: int64
      - name: lora_2_alpha
        dtype: float64
      - name: lora_3_name
        dtype: string
      - name: lora_3_A_weight
        sequence:
          sequence: float32
      - name: lora_3_B_weight
        sequence:
          sequence: float32
      - name: lora_3_rank
        dtype: int64
      - name: lora_3_alpha
        dtype: float64
      - name: lora_4_name
        dtype: string
      - name: lora_4_A_weight
        sequence:
          sequence: float32
      - name: lora_4_B_weight
        sequence:
          sequence: float32
      - name: lora_4_rank
        dtype: int64
      - name: lora_4_alpha
        dtype: float64
      - name: lora_5_name
        dtype: string
      - name: lora_5_A_weight
        sequence:
          sequence: float32
      - name: lora_5_B_weight
        sequence:
          sequence: float32
      - name: lora_5_rank
        dtype: int64
      - name: lora_5_alpha
        dtype: float64
      - name: lora_6_name
        dtype: string
      - name: lora_6_A_weight
        sequence:
          sequence: float32
      - name: lora_6_B_weight
        sequence:
          sequence: float32
      - name: lora_6_rank
        dtype: int64
      - name: lora_6_alpha
        dtype: float64
      - name: lora_7_name
        dtype: string
      - name: lora_7_A_weight
        sequence:
          sequence: float32
      - name: lora_7_B_weight
        sequence:
          sequence: float32
      - name: lora_7_rank
        dtype: int64
      - name: lora_7_alpha
        dtype: float64
      - name: lora_8_name
        dtype: string
      - name: lora_8_A_weight
        sequence:
          sequence: float32
      - name: lora_8_B_weight
        sequence:
          sequence: float32
      - name: lora_8_rank
        dtype: int64
      - name: lora_8_alpha
        dtype: float64
      - name: lora_9_name
        dtype: string
      - name: lora_9_A_weight
        sequence:
          sequence: float32
      - name: lora_9_B_weight
        sequence:
          sequence: float32
      - name: lora_9_rank
        dtype: int64
      - name: lora_9_alpha
        dtype: float64
      - name: lora_10_name
        dtype: string
      - name: lora_10_A_weight
        sequence:
          sequence: float32
      - name: lora_10_B_weight
        sequence:
          sequence: float32
      - name: lora_10_rank
        dtype: int64
      - name: lora_10_alpha
        dtype: float64
      - name: lora_11_name
        dtype: string
      - name: lora_11_A_weight
        sequence:
          sequence: float32
      - name: lora_11_B_weight
        sequence:
          sequence: float32
      - name: lora_11_rank
        dtype: int64
      - name: lora_11_alpha
        dtype: float64
      - name: lora_12_name
        dtype: string
      - name: lora_12_A_weight
        sequence:
          sequence: float32
      - name: lora_12_B_weight
        sequence:
          sequence: float32
      - name: lora_12_rank
        dtype: int64
      - name: lora_12_alpha
        dtype: float64
      - name: lora_13_name
        dtype: string
      - name: lora_13_A_weight
        sequence:
          sequence: float32
      - name: lora_13_B_weight
        sequence:
          sequence: float32
      - name: lora_13_rank
        dtype: int64
      - name: lora_13_alpha
        dtype: float64
      - name: lora_14_name
        dtype: string
      - name: lora_14_A_weight
        sequence:
          sequence: float32
      - name: lora_14_B_weight
        sequence:
          sequence: float32
      - name: lora_14_rank
        dtype: int64
      - name: lora_14_alpha
        dtype: float64
    splits:
      - name: train
        num_bytes: 2561357508
        num_examples: 264
    download_size: 1724766354
    dataset_size: 2561357508
  - config_name: vit
    features:
      - name: task_name
        dtype: string
      - name: layer_model
        dtype: string
      - name: layer_name
        dtype: string
      - name: pre_ft_name
        dtype: string
      - name: pre_ft_weight
        sequence:
          sequence: float32
      - name: lora_0_name
        dtype: string
      - name: lora_0_A_weight
        sequence:
          sequence: float32
      - name: lora_0_B_weight
        sequence:
          sequence: float32
      - name: lora_0_rank
        dtype: int64
      - name: lora_0_alpha
        dtype: int64
      - name: lora_1_name
        dtype: string
      - name: lora_1_A_weight
        sequence:
          sequence: float32
      - name: lora_1_B_weight
        sequence:
          sequence: float32
      - name: lora_1_rank
        dtype: int64
      - name: lora_1_alpha
        dtype: int64
      - name: lora_2_name
        dtype: string
      - name: lora_2_A_weight
        sequence:
          sequence: float32
      - name: lora_2_B_weight
        sequence:
          sequence: float32
      - name: lora_2_rank
        dtype: int64
      - name: lora_2_alpha
        dtype: int64
      - name: lora_3_name
        dtype: string
      - name: lora_3_A_weight
        sequence:
          sequence: float32
      - name: lora_3_B_weight
        sequence:
          sequence: float32
      - name: lora_3_rank
        dtype: int64
      - name: lora_3_alpha
        dtype: int64
      - name: lora_4_name
        dtype: string
      - name: lora_4_A_weight
        sequence:
          sequence: float32
      - name: lora_4_B_weight
        sequence:
          sequence: float32
      - name: lora_4_rank
        dtype: int64
      - name: lora_4_alpha
        dtype: int64
      - name: lora_5_name
        dtype: string
      - name: lora_5_A_weight
        sequence:
          sequence: float32
      - name: lora_5_B_weight
        sequence:
          sequence: float32
      - name: lora_5_rank
        dtype: int64
      - name: lora_5_alpha
        dtype: int64
      - name: lora_6_name
        dtype: string
      - name: lora_6_A_weight
        sequence:
          sequence: float32
      - name: lora_6_B_weight
        sequence:
          sequence: float32
      - name: lora_6_rank
        dtype: int64
      - name: lora_6_alpha
        dtype: int64
      - name: lora_7_name
        dtype: string
      - name: lora_7_A_weight
        sequence:
          sequence: float32
      - name: lora_7_B_weight
        sequence:
          sequence: float32
      - name: lora_7_rank
        dtype: int64
      - name: lora_7_alpha
        dtype: int64
      - name: lora_8_name
        dtype: string
      - name: lora_8_A_weight
        sequence:
          sequence: float32
      - name: lora_8_B_weight
        sequence:
          sequence: float32
      - name: lora_8_rank
        dtype: int64
      - name: lora_8_alpha
        dtype: int64
      - name: lora_9_name
        dtype: string
      - name: lora_9_A_weight
        sequence:
          sequence: float32
      - name: lora_9_B_weight
        sequence:
          sequence: float32
      - name: lora_9_rank
        dtype: int64
      - name: lora_9_alpha
        dtype: int64
      - name: lora_10_name
        dtype: string
      - name: lora_10_A_weight
        sequence:
          sequence: float32
      - name: lora_10_B_weight
        sequence:
          sequence: float32
      - name: lora_10_rank
        dtype: int64
      - name: lora_10_alpha
        dtype: int64
      - name: lora_11_name
        dtype: string
      - name: lora_11_A_weight
        sequence:
          sequence: float32
      - name: lora_11_B_weight
        sequence:
          sequence: float32
      - name: lora_11_rank
        dtype: int64
      - name: lora_11_alpha
        dtype: int64
      - name: lora_12_name
        dtype: string
      - name: lora_12_A_weight
        sequence:
          sequence: float32
      - name: lora_12_B_weight
        sequence:
          sequence: float32
      - name: lora_12_rank
        dtype: int64
      - name: lora_12_alpha
        dtype: int64
      - name: lora_13_name
        dtype: string
      - name: lora_13_A_weight
        sequence:
          sequence: float32
      - name: lora_13_B_weight
        sequence:
          sequence: float32
      - name: lora_13_rank
        dtype: int64
      - name: lora_13_alpha
        dtype: int64
      - name: lora_14_name
        dtype: string
      - name: lora_14_A_weight
        sequence:
          sequence: float32
      - name: lora_14_B_weight
        sequence:
          sequence: float32
      - name: lora_14_rank
        dtype: int64
      - name: lora_14_alpha
        dtype: int64
    splits:
      - name: train
        num_bytes: 93231628
        num_examples: 24
    download_size: 111481540
    dataset_size: 93231628
configs:
  - config_name: mistral-7b-v0.1-dpo
    data_files:
      - split: train
        path: mistral-7b-v0.1-dpo/train-*
  - config_name: mistral-7b-v0.1-sft
    data_files:
      - split: train
        path: mistral-7b-v0.1-sft/train-*
  - config_name: stable-diffusion-1.5
    data_files:
      - split: train
        path: stable-diffusion-1.5/train-*
  - config_name: vit
    data_files:
      - split: train
        path: vit/train-*