Update README.md
Browse files
README.md
CHANGED
|
@@ -238,15 +238,28 @@ print(expected_vals)
|
|
| 238 |
# Evaluation
|
| 239 |
|
| 240 |
We perform an evaluation on 9 datasets from the [BEIR benchmark](https://github.com/beir-cellar/beir) that none of the evaluated models have been trained upon (to our knowledge).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 241 |
We evaluate on a subset of all queries (the first 250) to save evaluation time.
|
| 242 |
|
| 243 |
We find that our model performs similarly or better than many of the state-of-the-art reranker models in our evaluation, without compromising on inference speed.
|
| 244 |
|
| 245 |
We make our evaluation code and results available [on our Github](https://github.com/lightblue-tech/lb-reranker/blob/main/run_bier.ipynb).
|
| 246 |
|
| 247 |
-
 that none of the evaluated models have been trained upon (to our knowledge).
|
| 241 |
+
|
| 242 |
+
* Arguana
|
| 243 |
+
* Dbpedia-entity
|
| 244 |
+
* Fiqa
|
| 245 |
+
* NFcorpus
|
| 246 |
+
* Scidocs
|
| 247 |
+
* Scifact
|
| 248 |
+
* Trec-covid-v2
|
| 249 |
+
* Vihealthqa
|
| 250 |
+
* Webis-touche2020
|
| 251 |
+
|
| 252 |
We evaluate on a subset of all queries (the first 250) to save evaluation time.
|
| 253 |
|
| 254 |
We find that our model performs similarly or better than many of the state-of-the-art reranker models in our evaluation, without compromising on inference speed.
|
| 255 |
|
| 256 |
We make our evaluation code and results available [on our Github](https://github.com/lightblue-tech/lb-reranker/blob/main/run_bier.ipynb).
|
| 257 |
|
| 258 |
+

|
| 259 |
+
|
| 260 |
+

|
| 261 |
|
| 262 |
+

|
| 263 |
|
| 264 |
# License
|
| 265 |
|