AkimfromParis commited on
Commit
882f81a
·
verified ·
1 Parent(s): c8e96bf

Update logos number 5

Browse files
Files changed (1) hide show
  1. src/about.py +4 -4
src/about.py CHANGED
@@ -75,13 +75,13 @@ TITLE = """<h1 align="center" id="space-title">🇯🇵Open Japanese LLM Leaderb
75
  BOTTOM_LOGO = """
76
  <div style="display: flex; flex-direction: row; justify-content: space-around; align-items: center" dir="ltr">
77
  <a href="https://llm-jp.nii.ac.jp/en/">
78
- <img src="file/Logos-HQ/LLM-jp-Logo-Oct-2024.png" alt="LLM-jp" style="max-height: 100px">
79
  </a>
80
  <a href="https://mdx.jp/">
81
- <img src="file/Logos-HQ/MDX-Logo-Oct-2024.jpg" alt="MDX platform" style="max-height: 100px">
82
  </a>
83
  <a href="https://huggingface.co/">
84
- <img src="file/Logos-HQ/HuggingFace-Logo-Oct-2024.png" alt="Hugging Face" style="max-height: 100px">
85
  </a>
86
  </div>
87
  </div>
@@ -97,7 +97,7 @@ On the __"LLM Benchmark"__ page, the question mark **"?"** refers to the paramet
97
  # Which evaluations are you running? how can people reproduce what you have?
98
  LLM_BENCHMARKS_TEXT = f"""
99
  ## How it works
100
- 📈 We evaluate Japanese Large Language Models on 51 key benchmarks leveraging our evaluation tool [llm-jp-eval](https://github.com/llm-jp/llm-jp-eval), a unified framework to evaluate Japanese LLMs on various evaluation tasks.
101
 
102
  **NLI (Natural Language Inference)**
103
 
 
75
  BOTTOM_LOGO = """
76
  <div style="display: flex; flex-direction: row; justify-content: space-around; align-items: center" dir="ltr">
77
  <a href="https://llm-jp.nii.ac.jp/en/">
78
+ <img src="file/src/Logos-HQ/LLM-jp-Logo-Oct-2024.png" alt="LLM-jp" style="max-height: 100px">
79
  </a>
80
  <a href="https://mdx.jp/">
81
+ <img src="file/src/Logos-HQ/MDX-Logo-Oct-2024.jpg" alt="MDX platform" style="max-height: 100px">
82
  </a>
83
  <a href="https://huggingface.co/">
84
+ <img src="file/src/Logos-HQ/HuggingFace-Logo-Oct-2024.png" alt="Hugging Face" style="max-height: 100px">
85
  </a>
86
  </div>
87
  </div>
 
97
  # Which evaluations are you running? how can people reproduce what you have?
98
  LLM_BENCHMARKS_TEXT = f"""
99
  ## How it works
100
+ 📈 We evaluate Japanese Large Language Models on 16 tasks leveraging our evaluation tool [llm-jp-eval](https://github.com/llm-jp/llm-jp-eval), a unified framework to evaluate Japanese LLMs on various evaluation tasks.
101
 
102
  **NLI (Natural Language Inference)**
103