ldwang commited on
Commit
a268908
β€’
1 Parent(s): 2f68ed7

update readme

Browse files
Files changed (1) hide show
  1. README.md +3 -28
README.md CHANGED
@@ -2604,20 +2604,7 @@ pipeline_tag: sentence-similarity
2604
 
2605
 
2606
  <h1 align="center">FlagEmbedding</h1>
2607
- <p align="center">
2608
- <a href="https://www.python.org/">
2609
- <img alt="Build" src="https://img.shields.io/badge/Contribution-Welcome-blue">
2610
- </a>
2611
- <a href="https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE">
2612
- <img alt="License" src="https://img.shields.io/badge/LICENSE-MIT-green">
2613
- </a>
2614
- <a href="https://huggingface.co/C-MTEB">
2615
- <img alt="Build" src="https://img.shields.io/badge/C_MTEB-πŸ€—-yellow">
2616
- </a>
2617
- <a href="https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding">
2618
- <img alt="Build" src="https://img.shields.io/badge/FlagEmbedding-1.0-red">
2619
- </a>
2620
- </p>
2621
 
2622
  <h4 align="center">
2623
  <p>
@@ -2630,6 +2617,7 @@ pipeline_tag: sentence-similarity
2630
  <p>
2631
  </h4>
2632
 
 
2633
 
2634
  [English](README.md) | [δΈ­ζ–‡](https://github.com/FlagOpen/FlagEmbedding/blob/master/README_zh.md)
2635
 
@@ -2637,7 +2625,7 @@ FlagEmbedding can map any text to a low-dimensional dense vector which can be us
2637
  And it also can be used in vector database for LLMs.
2638
 
2639
  ************* 🌟**Updates**🌟 *************
2640
- - 08/09/2023: BGE Models are integrated into Langchain, you can use it like [this](#using-langchain); C-MTEB leaderboard is [avaliable](https://huggingface.co/spaces/mteb/leaderboard).
2641
  - 08/05/2023: Release base-scale and small-scale models, **best performance among the models of the same size πŸ€—**
2642
  - 08/02/2023: Release `bge-large-*`(short for BAAI General Embedding) Models, **rank 1st on MTEB and C-MTEB benchmark!** :tada: :tada:
2643
  - 08/01/2023: We release the [Chinese Massive Text Embedding Benchmark](https://github.com/FlagOpen/FlagEmbedding/blob/master/C_MTEB) (**C-MTEB**), consisting of 31 test dataset.
@@ -2867,23 +2855,10 @@ You can easily finetune your model with it.
2867
 
2868
  **The data collection is to be released in the future.**
2869
 
2870
-
2871
- ## Schedule
2872
- - [x] Chinese Massive Text Embedding Benchmark
2873
- - [x] release baai-general-embedding models
2874
- - [x] release codes for training
2875
- - [ ] Multilingual model
2876
- - [ ] Training Datasets
2877
- - [ ] ...
2878
-
2879
  We will continually update the embedding models and training codes,
2880
  hoping to promote the development of the embedding model community.
2881
 
2882
 
2883
- ## Contact
2884
- If you have any question or suggestion related to this project, feel free to open an issue or pull a request.
2885
- You also can email Shitao Xiao(stxiao@baai.ac.cn) and Zheng Liu(liuzheng@baai.ac.cn).
2886
-
2887
 
2888
  ## License
2889
  FlagEmbedding is licensed under [MIT License](https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE). The released models can be used for commercial purposes free of charge.
 
2604
 
2605
 
2606
  <h1 align="center">FlagEmbedding</h1>
2607
+
 
 
 
 
 
 
 
 
 
 
 
 
 
2608
 
2609
  <h4 align="center">
2610
  <p>
 
2617
  <p>
2618
  </h4>
2619
 
2620
+ More details please refer to our Github: [FlagEmbedding](https://github.com/FlagOpen/FlagEmbedding).
2621
 
2622
  [English](README.md) | [δΈ­ζ–‡](https://github.com/FlagOpen/FlagEmbedding/blob/master/README_zh.md)
2623
 
 
2625
  And it also can be used in vector database for LLMs.
2626
 
2627
  ************* 🌟**Updates**🌟 *************
2628
+ - 08/09/2023: BGE Models are integrated into **Langchain**, you can use it like [this](#using-langchain); C-MTEB **leaderboard** is [avaliable](https://huggingface.co/spaces/mteb/leaderboard).
2629
  - 08/05/2023: Release base-scale and small-scale models, **best performance among the models of the same size πŸ€—**
2630
  - 08/02/2023: Release `bge-large-*`(short for BAAI General Embedding) Models, **rank 1st on MTEB and C-MTEB benchmark!** :tada: :tada:
2631
  - 08/01/2023: We release the [Chinese Massive Text Embedding Benchmark](https://github.com/FlagOpen/FlagEmbedding/blob/master/C_MTEB) (**C-MTEB**), consisting of 31 test dataset.
 
2855
 
2856
  **The data collection is to be released in the future.**
2857
 
 
 
 
 
 
 
 
 
 
2858
  We will continually update the embedding models and training codes,
2859
  hoping to promote the development of the embedding model community.
2860
 
2861
 
 
 
 
 
2862
 
2863
  ## License
2864
  FlagEmbedding is licensed under [MIT License](https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE). The released models can be used for commercial purposes free of charge.