Update readme
Browse files
README.md
CHANGED
|
@@ -29,7 +29,7 @@ tags:
|
|
| 29 |
|
| 30 |
**Youtu-Embedding** is a state-of-the-art, general-purpose text embedding model developed by Tencent Youtu Lab. It delivers exceptional performance across a wide range of natural language processing tasks, including Information Retrieval (IR), Semantic Textual Similarity (STS), Clustering, Reranking, and Classification.
|
| 31 |
|
| 32 |
-
- **Top-Ranked Performance**: Achieved the #1 score of **77.
|
| 33 |
|
| 34 |
- **Innovative Training Framework**: Features a Collaborative-Discriminative Fine-tuning Framework designed to resolve the "negative transfer" problem in multi-task learning. This is accomplished through a unified data format, task-differentiated loss functions, and a dynamic single-task sampling mechanism.
|
| 35 |
|
|
@@ -48,13 +48,13 @@ tags:
|
|
| 48 |
#### 1. Using `transformers`
|
| 49 |
**📦 Installation**
|
| 50 |
```bash
|
| 51 |
-
pip install transformers==4.51.3
|
| 52 |
```
|
| 53 |
**⚙️ Usage**
|
| 54 |
```python
|
| 55 |
import torch
|
| 56 |
import numpy as np
|
| 57 |
-
from transformers import AutoModel,
|
| 58 |
|
| 59 |
|
| 60 |
class LLMEmbeddingModel():
|
|
|
|
| 29 |
|
| 30 |
**Youtu-Embedding** is a state-of-the-art, general-purpose text embedding model developed by Tencent Youtu Lab. It delivers exceptional performance across a wide range of natural language processing tasks, including Information Retrieval (IR), Semantic Textual Similarity (STS), Clustering, Reranking, and Classification.
|
| 31 |
|
| 32 |
+
- **Top-Ranked Performance**: Achieved the #1 score of **77.58** on the authoritative CMTEB (Chinese Massive Text Embedding Benchmark) as of September 2025, demonstrating its powerful and robust text representation capabilities.
|
| 33 |
|
| 34 |
- **Innovative Training Framework**: Features a Collaborative-Discriminative Fine-tuning Framework designed to resolve the "negative transfer" problem in multi-task learning. This is accomplished through a unified data format, task-differentiated loss functions, and a dynamic single-task sampling mechanism.
|
| 35 |
|
|
|
|
| 48 |
#### 1. Using `transformers`
|
| 49 |
**📦 Installation**
|
| 50 |
```bash
|
| 51 |
+
pip install transformers==4.51.3
|
| 52 |
```
|
| 53 |
**⚙️ Usage**
|
| 54 |
```python
|
| 55 |
import torch
|
| 56 |
import numpy as np
|
| 57 |
+
from transformers import AutoModel, AutoTokenizer
|
| 58 |
|
| 59 |
|
| 60 |
class LLMEmbeddingModel():
|