update readme
Browse files
README.md
CHANGED
@@ -2604,20 +2604,7 @@ pipeline_tag: sentence-similarity
|
|
2604 |
|
2605 |
|
2606 |
<h1 align="center">FlagEmbedding</h1>
|
2607 |
-
|
2608 |
-
<a href="https://www.python.org/">
|
2609 |
-
<img alt="Build" src="https://img.shields.io/badge/Contribution-Welcome-blue">
|
2610 |
-
</a>
|
2611 |
-
<a href="https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE">
|
2612 |
-
<img alt="License" src="https://img.shields.io/badge/LICENSE-MIT-green">
|
2613 |
-
</a>
|
2614 |
-
<a href="https://huggingface.co/C-MTEB">
|
2615 |
-
<img alt="Build" src="https://img.shields.io/badge/C_MTEB-π€-yellow">
|
2616 |
-
</a>
|
2617 |
-
<a href="https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding">
|
2618 |
-
<img alt="Build" src="https://img.shields.io/badge/FlagEmbedding-1.0-red">
|
2619 |
-
</a>
|
2620 |
-
</p>
|
2621 |
|
2622 |
<h4 align="center">
|
2623 |
<p>
|
@@ -2630,6 +2617,7 @@ pipeline_tag: sentence-similarity
|
|
2630 |
<p>
|
2631 |
</h4>
|
2632 |
|
|
|
2633 |
|
2634 |
[English](README.md) | [δΈζ](https://github.com/FlagOpen/FlagEmbedding/blob/master/README_zh.md)
|
2635 |
|
@@ -2637,7 +2625,7 @@ FlagEmbedding can map any text to a low-dimensional dense vector which can be us
|
|
2637 |
And it also can be used in vector database for LLMs.
|
2638 |
|
2639 |
************* π**Updates**π *************
|
2640 |
-
- 08/09/2023: BGE Models are integrated into Langchain
|
2641 |
- 08/05/2023: Release base-scale and small-scale models, **best performance among the models of the same size π€**
|
2642 |
- 08/02/2023: Release `bge-large-*`(short for BAAI General Embedding) Models, **rank 1st on MTEB and C-MTEB benchmark!** :tada: :tada:
|
2643 |
- 08/01/2023: We release the [Chinese Massive Text Embedding Benchmark](https://github.com/FlagOpen/FlagEmbedding/blob/master/C_MTEB) (**C-MTEB**), consisting of 31 test dataset.
|
@@ -2867,23 +2855,10 @@ You can easily finetune your model with it.
|
|
2867 |
|
2868 |
**The data collection is to be released in the future.**
|
2869 |
|
2870 |
-
|
2871 |
-
## Schedule
|
2872 |
-
- [x] Chinese Massive Text Embedding Benchmark
|
2873 |
-
- [x] release baai-general-embedding models
|
2874 |
-
- [x] release codes for training
|
2875 |
-
- [ ] Multilingual model
|
2876 |
-
- [ ] Training Datasets
|
2877 |
-
- [ ] ...
|
2878 |
-
|
2879 |
We will continually update the embedding models and training codes,
|
2880 |
hoping to promote the development of the embedding model community.
|
2881 |
|
2882 |
|
2883 |
-
## Contact
|
2884 |
-
If you have any question or suggestion related to this project, feel free to open an issue or pull a request.
|
2885 |
-
You also can email Shitao Xiao(stxiao@baai.ac.cn) and Zheng Liu(liuzheng@baai.ac.cn).
|
2886 |
-
|
2887 |
|
2888 |
## License
|
2889 |
FlagEmbedding is licensed under [MIT License](https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE). The released models can be used for commercial purposes free of charge.
|
|
|
2604 |
|
2605 |
|
2606 |
<h1 align="center">FlagEmbedding</h1>
|
2607 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2608 |
|
2609 |
<h4 align="center">
|
2610 |
<p>
|
|
|
2617 |
<p>
|
2618 |
</h4>
|
2619 |
|
2620 |
+
More details please refer to our Github: [FlagEmbedding](https://github.com/FlagOpen/FlagEmbedding).
|
2621 |
|
2622 |
[English](README.md) | [δΈζ](https://github.com/FlagOpen/FlagEmbedding/blob/master/README_zh.md)
|
2623 |
|
|
|
2625 |
And it also can be used in vector database for LLMs.
|
2626 |
|
2627 |
************* π**Updates**π *************
|
2628 |
+
- 08/09/2023: BGE Models are integrated into **Langchain**, you can use it like [this](#using-langchain); C-MTEB **leaderboard** is [avaliable](https://huggingface.co/spaces/mteb/leaderboard).
|
2629 |
- 08/05/2023: Release base-scale and small-scale models, **best performance among the models of the same size π€**
|
2630 |
- 08/02/2023: Release `bge-large-*`(short for BAAI General Embedding) Models, **rank 1st on MTEB and C-MTEB benchmark!** :tada: :tada:
|
2631 |
- 08/01/2023: We release the [Chinese Massive Text Embedding Benchmark](https://github.com/FlagOpen/FlagEmbedding/blob/master/C_MTEB) (**C-MTEB**), consisting of 31 test dataset.
|
|
|
2855 |
|
2856 |
**The data collection is to be released in the future.**
|
2857 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2858 |
We will continually update the embedding models and training codes,
|
2859 |
hoping to promote the development of the embedding model community.
|
2860 |
|
2861 |
|
|
|
|
|
|
|
|
|
2862 |
|
2863 |
## License
|
2864 |
FlagEmbedding is licensed under [MIT License](https://github.com/FlagOpen/FlagEmbedding/blob/master/LICENSE). The released models can be used for commercial purposes free of charge.
|