You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/evaluation/README.md
+26-22Lines changed: 26 additions & 22 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -54,47 +54,29 @@ First, we will introduce the commonly used parameters, followed by an introducti
54
54
**Parameters for Model Configuration:**
55
55
56
56
-**`embedder_name_or_path`**: The name or path to the embedder.
57
-
58
57
-**`embedder_model_class`**: Class of the model used for embedding (options include 'auto', 'encoder-only-base', etc.). Default is `auto`.
59
-
60
58
-**`normalize_embeddings`**: Set to `true` to normalize embeddings.
61
-
62
59
-**`use_fp16`**: Use FP16 precision for inference.
63
-
64
60
-**`devices`**: List of devices used for inference.
65
-
66
61
-**`query_instruction_for_retrieval`**, **`query_instruction_format_for_retrieval`**: Instructions and format for query during retrieval.
67
-
68
62
-**`examples_for_task`**, **`examples_instruction_format`**: Example tasks and their instructions format.
69
-
70
63
-**`trust_remote_code`**: Set to `true` to trust remote code execution.
71
-
72
64
-**`reranker_name_or_path`**: Name or path to the reranker.
73
-
74
65
-**`reranker_model_class`**: Reranker model class (options include 'auto', 'decoder-only-base', etc.). Default is `auto`.
75
-
76
66
-**`reranker_peft_path`**: Path for portable encoder fine-tuning of the reranker.
77
-
78
67
-**`use_bf16`**: Use BF16 precision for inference.
79
-
80
68
-**`query_instruction_for_rerank`**, **`query_instruction_format_for_rerank`**: Instructions and format for query during reranking.
81
-
82
69
-**`passage_instruction_for_rerank`**, **`passage_instruction_format_for_rerank`**: Instructions and format for processing passages during reranking.
83
-
84
70
-**`cache_dir`**: Cache directory for models.
85
-
86
71
-**`embedder_batch_size`**, **`reranker_batch_size`**: Batch sizes for embedding and reranking.
87
-
88
72
-**`embedder_query_max_length`**, **`embedder_passage_max_length`**: Maximum length for embedding queries and passages.
89
-
90
73
-**`reranker_query_max_length`**, **`reranker_max_length`**: Maximum lengths for reranking queries and reranking in general.
91
-
92
74
-**`normalize`**: Normalize the reranking scores.
93
-
94
75
-**`prompt`**: Prompt for the reranker.
95
-
96
76
-**`cutoff_layers`**, **`compress_ratio`**, **`compress_layers`**: Parameters for configuring the output and compression of layerwise or lightweight rerankers.
97
77
78
+
***Notice:*** If you evaluate your own model, please set `embedder_model_class` and `reranker_model_class`.
You can refer to [MLDR dataset](https://github.com/hanhainebula/FlagEmbedding/tree/new-flagembedding-v1/FlagEmbedding/evaluation/mldr), just need to rewrite `DataLoader`, rewriting the loading method for the required dataset.
310
-
311
291
The example data for `corpus.jsonl`:
312
292
313
293
```json
@@ -334,3 +314,27 @@ The example data for `test_qrels.jsonl`:
0 commit comments