Interesting, but this aspect makes me double-take: "We demonstrate that Mistral-7B, when fine-tuned solely on synthetic data, attains competitive
performance on the BEIR [ 40 ] and MTEB [27] benchmarks".
E5/BGE large are an order of magnitude smaller than Mistral-7B. So is this just "bigger model wins" in disguise?
I need to read the whole paper carefully, but this jumped out at me.
Interesting, but this aspect makes me double-take: "We demonstrate that Mistral-7B, when fine-tuned solely on synthetic data, attains competitive performance on the BEIR [ 40 ] and MTEB [27] benchmarks".
E5/BGE large are an order of magnitude smaller than Mistral-7B. So is this just "bigger model wins" in disguise?
I need to read the whole paper carefully, but this jumped out at me.