Feature Extraction
sentence-transformers
PyTorch
Safetensors
Transformers
English
mistral
mteb
Eval Results (legacy)
Eval Results
text-embeddings-inference
Instructions to use intfloat/e5-mistral-7b-instruct with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use intfloat/e5-mistral-7b-instruct with sentence-transformers:
from sentence_transformers import SentenceTransformer model = SentenceTransformer("intfloat/e5-mistral-7b-instruct") sentences = [ "The weather is lovely today.", "It's so sunny outside!", "He drove to the stadium." ] embeddings = model.encode(sentences) similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] - Transformers
How to use intfloat/e5-mistral-7b-instruct with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("feature-extraction", model="intfloat/e5-mistral-7b-instruct")# Load model directly from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("intfloat/e5-mistral-7b-instruct") model = AutoModel.from_pretrained("intfloat/e5-mistral-7b-instruct") - Inference
- Notebooks
- Google Colab
- Kaggle
Question about MTEB Evaluation and max_seq_length Settings for e5-mistral-7b-instruct
#47
by george31 - opened
I encountered an OOM error (using 96GB GPU) while running MTEB evaluation (on Miracl dataset)on e5-mistral-7b without specifying max_seq_length. While I managed to run the evaluation by reducing the max_seq_length, I have some concerns about the proper way to conduct these evaluations.
Current situation:
- Model: e5-instruct-7b
- Issue: OOM error when max_seq_length is not explicitly set
- Hardware: 96GB GPU memory
Questions:
- What is the recommended approach for setting max_seq_length when running MTEB evaluations, especially for large language models?
- Is there an industry standard or best practice for determining max_seq_length in benchmark evaluations?
- If we need to limit max_seq_length due to hardware constraints, how do we ensure fair comparison with other models in the leaderboard?
- Should we explicitly mention the max_seq_length used in our evaluation when reporting results?
I'd appreciate any insights from the community on handling sequence length limitations during benchmark evaluations, especially for resource-intensive models.