File size: 3,978 Bytes
6a47e4e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94aa50d
6a47e4e
94aa50d
6a47e4e
 
 
 
 
 
 
 
94aa50d
6a47e4e
 
 
 
 
 
94aa50d
6a47e4e
 
 
94aa50d
6a47e4e
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
base_model:
- Alibaba-NLP/gte-Qwen2-7B-instruct
language:
- en
- zh
tags:
- sentence-similarity
- transformers
---

# INF-Retriever-Qwen2-7B-Preview

- **INF-Retriever-Qwen2-7B-Preview** is a temporary version of the **INF-Retriever-Qwen2-7B** model, which has been fine-tuned for dense retrieval tasks in both Chinese and English. This preview version is being made available for testing and evaluation purposes.

- **Please note:** The **INF-Retriever-Qwen2-7B** model, which includes further refinements and optimizations, will be released as the official version soon. We encourage you to stay tuned for the upcoming full release, which will offer enhanced performance and additional features.

## Model Overview

- **INF-Retriever-Qwen2-7B** is an LLM-based dense retrieval model developed by [INF TECH](https://www.infly.cn/en). It is built upon the [gte-Qwen2-7B-instruct](https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct) model and specifically fine-tuned to excel in retrieval tasks, particularly for Chinese and English data.

- As of November 13, 2024, **INF-Retriever-Qwen2-7B-Preview** ranks **No.1** on the Automated Heterogeneous Information Retrieval Benchmark ([AIR-Bench](https://huggingface.co/spaces/AIR-Bench/leaderboard)), showcasing its cutting-edge performance in heterogeneous information retrieval tasks.

## Key Features

- **Optimized for Chinese and English retrieval**: The model has been specifically fine-tuned with retrieval-focused datasets in both languages, significantly improving its accuracy and efficiency for a variety of retrieval scenarios.
- **Top-tier performance**: **INF-Retriever-Qwen2-7B** has achieved outstanding results on the AIR-Bench leaderboard, making it a top choice for heterogeneous information retrieval tasks across various domains.

## AIR-Bench Evaluation Results

**INF-Retriever-Qwen2-7B-Preview** has demonstrated superior retrieval capabilities across multiple domains and languages. The results from the Automated Heterogeneous Information Retrieval Benchmark (AIR-Bench) as of November 13 are as follows:

|                                    Model Name                                     |  Average  |  wiki_en  |  wiki_zh  |  web_en   |  web_zh   | healthcare_en | healthcare_zh |  law_en   | arxiv_en  | news_en   | news_zh   | finance_en | finance_zh | msmarco_en | 
|:---------------------------------------------------------------------------------:|:---------:|:---------:|:---------:|:---------:|:---------:|:-------------:|:-------------:|:---------:|:---------:|-----------|-----------|------------|------------|------------|
|                   [BGE-M3](https://huggingface.co/BAAI/bge-m3)                    |   46.65   |   60.49   |   62.36   |   47.35   |   50.38   |     49.1      |   **42.38**   |   26.68   |   40.76   | 48.04     | 40.75     | 51.52      | 32.18      | 54.4       |
|  [BGE-Multilingual-Gemma2](https://huggingface.co/BAAI/bge-multilingual-gemma2)   |   46.83   |   63.71   |   67.3    |   50.38   |   53.24   |     47.24     |     42.13     |   22.58   |   23.28   | 50.91     | 44.02     | 49.3       | 31.6       | **63.14**  |
| [GTE-Qwen2-7B-instruct](https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct) |   48.38   |   63.46   |   66.44   |   51.2    |   51.98   |     54.2      |     38.82     |   22.31   |   40.27   | **54.07** | 43.03     | 58.2       | 26.63      | 58.39      |
|                            **INF-Retriever-Qwen2-7B-Preview**                             | **52.22** | **64.96** | **67.87** | **52.84** | **55.54** |   **58.82**   |     37.71     | **34.89** | **52.35** | 53.6      | **47.93** | **58.5**   | **33.92**  | 59.96      |

## Final Release Coming Soon

The official version of **INF-Retriever-Qwen2-7B** will provide even better performance, stability, and additional features. We are working hard to finalize the model, and we look forward to sharing the full release with the community in the near future. 

---