|
node_lines: |
|
- node_line_name: retrieve_node_line |
|
nodes: |
|
- node_type: retrieval |
|
strategy: |
|
metrics: [ retrieval_f1, retrieval_recall, retrieval_precision, |
|
retrieval_ndcg, retrieval_map, retrieval_mrr ] |
|
speed_threshold: 10 |
|
top_k: 10 |
|
modules: |
|
- module_type: bm25 |
|
bm25_tokenizer: [ porter_stemmer, space, gpt2 ] |
|
- module_type: vectordb |
|
embedding_model: openai |
|
embedding_batch: 256 |
|
- module_type: hybrid_rrf |
|
weight_range: (4,80) |
|
- module_type: hybrid_cc |
|
normalize_method: [ mm, tmm, z, dbsf ] |
|
weight_range: (0.0, 1.0) |
|
test_weight_size: 101 |
|
- node_type: passage_augmenter |
|
strategy: |
|
metrics: [ retrieval_f1, retrieval_recall, retrieval_precision ] |
|
speed_threshold: 5 |
|
top_k: 5 |
|
embedding_model: openai |
|
modules: |
|
- module_type: pass_passage_augmenter |
|
- module_type: prev_next_augmenter |
|
mode: next |
|
- node_type: passage_reranker |
|
strategy: |
|
metrics: [retrieval_f1, retrieval_recall, retrieval_precision] |
|
speed_threshold: 10 |
|
top_k: 5 |
|
modules: |
|
- module_type: pass_reranker |
|
- module_type: tart |
|
- module_type: monot5 |
|
- module_type: upr |
|
- module_type: cohere_reranker |
|
- module_type: rankgpt |
|
- module_type: jina_reranker |
|
- module_type: colbert_reranker |
|
- module_type: sentence_transformer_reranker |
|
- module_type: flag_embedding_reranker |
|
- module_type: flag_embedding_llm_reranker |
|
- module_type: time_reranker |
|
- module_type: openvino_reranker |
|
- module_type: voyageai_reranker |
|
- module_type: mixedbreadai_reranker |
|
- node_type: passage_filter |
|
strategy: |
|
metrics: [ retrieval_f1, retrieval_recall, retrieval_precision ] |
|
speed_threshold: 5 |
|
modules: |
|
- module_type: pass_passage_filter |
|
- module_type: similarity_threshold_cutoff |
|
threshold: 0.85 |
|
- module_type: similarity_percentile_cutoff |
|
percentile: 0.6 |
|
- module_type: threshold_cutoff |
|
threshold: 0.85 |
|
- module_type: percentile_cutoff |
|
percentile: 0.6 |
|
- node_line_name: post_retrieve_node_line |
|
nodes: |
|
- node_type: prompt_maker |
|
strategy: |
|
metrics: |
|
- metric_name: bleu |
|
- metric_name: meteor |
|
- metric_name: rouge |
|
- metric_name: sem_score |
|
embedding_model: openai |
|
speed_threshold: 10 |
|
generator_modules: |
|
- module_type: llama_index_llm |
|
llm: openai |
|
model: [gpt-4o-mini] |
|
modules: |
|
- module_type: fstring |
|
prompt: ["Tell me something about the question: {query} \n\n {retrieved_contents}", |
|
"Question: {query} \n Something to read: {retrieved_contents} \n What's your answer?"] |
|
- module_type: long_context_reorder |
|
prompt: [ "Tell me something about the question: {query} \n\n {retrieved_contents}", |
|
"Question: {query} \n Something to read: {retrieved_contents} \n What's your answer?" ] |
|
- node_type: generator |
|
strategy: |
|
metrics: |
|
- metric_name: bleu |
|
- metric_name: meteor |
|
- metric_name: rouge |
|
- metric_name: sem_score |
|
embedding_model: openai |
|
speed_threshold: 10 |
|
modules: |
|
- module_type: llama_index_llm |
|
llm: [openai] |
|
model: [gpt-4o-mini] |
|
temperature: [0.5, 1.0] |
|
|