You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- 7/26/2024: Release a new embedding model [bge-en-icl](BAAI/bge-en-icl), an embedding model that incorporates in-context learning capabilities, which, by providing task-relevant query-response examples, can encode semantically richer queries, further enhancing the semantic representation ability of the embeddings.
41
+
- 7/26/2024: Release a new lightweight reranker [bge-reranker-v2.5-gemma2-lightweight](BAAI/bge-reranker-v2.5-gemma2-lightweight), a lightweight reranker based on gemma2-9B, which supports token compression and layerwise lightweight operations, can still ensure good performance while saving a significant amount of resources.
40
42
- 6/7/2024: Release a new benchmark [MLVU](https://github.com/JUNJIE99/MLVU), the first comprehensive benchmark specifically designed for long video understanding. MLVU features an extensive range of video durations, a diverse collection of video sources, and a set of evaluation tasks uniquely tailored for long-form video understanding. :fire:
41
43
- 5/21/2024: Release a new benchmark [AIR-Bench](https://github.com/AIR-Bench/AIR-Bench) together with Jina AI, Zilliz, HuggingFace, and other partners. AIR-Bench focuses on a fair out-of-distribution evaluation for Neural IR & RAG. It generates the synthetic data for benchmarking w.r.t. diverse domains and languages. It is dynamic and will be updated on regular basis. [Leaderboard](https://huggingface.co/spaces/AIR-Bench/leaderboard):fire:
42
44
- 4/30/2024: Release [Llama-3-8B-Instruct-80K-QLoRA](https://huggingface.co/namespace-Pt/Llama-3-8B-Instruct-80K-QLoRA), extending the context length of Llama-3-8B-Instruct from 8K to 80K via QLoRA training on a few synthesized long-context data. The model achieves remarkable performance on various long-context benchmarks. [Code](https://github.com/FlagOpen/FlagEmbedding/tree/master/Long_LLM/longllm_qlora):fire:
@@ -158,9 +160,14 @@ Refer to our [report: c-pack](https://arxiv.org/pdf/2309.07597.pdf) and [code](h
158
160
159
161
| Model | Language || Description | query instruction for retrieval |
|[BAAI/bge-en-icl](https://huggingface.co/BAAI/bge-en-icl)| English || A LLM-based embedding model with in-context learning capabilities, which can fully leverage the model's potential based on a few shot examples | Provide instructions and few-shot examples freely based on the given task. |
|[LM-Cocktail](https://huggingface.co/Shitao)| English || fine-tuned models (Llama and BGE) which can be used to reproduce the results of LM-Cocktail ||
163
166
|[BAAI/llm-embedder](https://huggingface.co/BAAI/llm-embedder)| English |[Inference](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_embedder)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_embedder)| a unified embedding model to support diverse retrieval augmentation needs for LLMs | See [README](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_embedder)|
167
+
|[BAAI/bge-reranker-v2-m3](https://huggingface.co/BAAI/bge-reranker-v2-m3)| Multilingual |[Inference](#usage-for-reranker)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker)| a lightweight cross-encoder model, possesses strong multilingual capabilities, easy to deploy, with fast inference. ||
168
+
|[BAAI/bge-reranker-v2-gemma](https://huggingface.co/BAAI/bge-reranker-v2-gemma)| Multilingual |[Inference](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker)| a cross-encoder model which is suitable for multilingual contexts, performs well in both English proficiency and multilingual capabilities. ||
169
+
|[BAAI/bge-reranker-v2-minicpm-layerwise](https://huggingface.co/BAAI/bge-reranker-v2-minicpm-layerwise)| Multilingual |[Inference](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker)| a cross-encoder model which is suitable for multilingual contexts, performs well in both English and Chinese proficiency, allows freedom to select layers for output, facilitating accelerated inference. ||
170
+
|[BAAI/bge-reranker-v2.5-gemma2-lightweight](https://huggingface.co/BAAI/bge-reranker-v2.5-gemma2-lightweight)| Multilingual |[Inference](BAAI/bge-reranker-v2.5-gemma2-lightweight)| a cross-encoder model which is suitable for multilingual contexts, performs well in both English and Chinese proficiency, allows freedom to select layers, compress ratio and compress layers for output, facilitating accelerated inference. ||
164
171
|[BAAI/bge-reranker-large](https://huggingface.co/BAAI/bge-reranker-large)| Chinese and English |[Inference](#usage-for-reranker)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker)| a cross-encoder model which is more accurate but less efficient ||
165
172
|[BAAI/bge-reranker-base](https://huggingface.co/BAAI/bge-reranker-base)| Chinese and English |[Inference](#usage-for-reranker)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker)| a cross-encoder model which is more accurate but less efficient ||
166
173
|[BAAI/bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5)| English |[Inference](#usage-for-embedding-model)[Fine-tune](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/finetune)| version 1.5 with more reasonable similarity distribution |`Represent this sentence for searching relevant passages: `|
|[BAAI/bge-reranker-large](https://huggingface.co/BAAI/bge-reranker-large)| Chinese and English |[推理](#usage-for-reranker)[微调](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker)| 交叉编码器模型,精度比向量模型更高但推理效率较低 [2]||
147
154
|[BAAI/bge-reranker-base](https://huggingface.co/BAAI/bge-reranker-base)| Chinese and English |[推理](#usage-for-reranker)[微调](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker)| 交叉编码器模型,精度比向量模型更高但推理效率较低 [2]||
148
155
|[BAAI/bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5)| English |[推理](#usage-for-embedding-model)[微调](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/finetune)| 1.5版本,相似度分布更加合理 |`Represent this sentence for searching relevant passages: `|
0 commit comments