Text Embedding Inference
This notebook demonstrates how to configure TextEmbeddingInference
embeddings.
The first step is to deploy the embeddings server. For detailed instructions, see the official repository for Text Embeddings Inference. Or tei-gaudi repository if you are deploying on Habana Gaudi/Gaudi 2.
Once deployed, the code below will connect to and submit embeddings for inference.
If youβre opening this Notebook on colab, you will probably need to install LlamaIndex π¦.
%pip install llama-index-embeddings-text-embeddings-inference
!pip install llama-index
from llama_index.embeddings.text_embeddings_inference import ( TextEmbeddingsInference,)
embed_model = TextEmbeddingsInference( model_name="BAAI/bge-large-en-v1.5", # required for formatting inference text, timeout=60, # timeout in seconds embed_batch_size=10, # batch size for embedding)
embeddings = embed_model.get_text_embedding("Hello World!")print(len(embeddings))print(embeddings[:5])
1024[0.010597229, 0.05895996, 0.022445679, -0.012046814, -0.03164673]
embeddings = await embed_model.aget_text_embedding("Hello World!")print(len(embeddings))print(embeddings[:5])
1024[0.010597229, 0.05895996, 0.022445679, -0.012046814, -0.03164673]