You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When embed_batch_size < len(texts), multiple EmbeddingEndEvent instances are emitted. However, the event handler in openinference.instrumentation.llama_index._handler.py processes these events by using keys like {EMBEDDING_EMBEDDINGS}.{i}.{EMBEDDING_TEXT} to store data:
Describe the Bug
In the
llamindex
BaseEmbedding class:When
embed_batch_size < len(texts)
, multipleEmbeddingEndEvent
instances are emitted. However, the event handler inopeninference.instrumentation.llama_index._handler.py
processes these events by using keys like{EMBEDDING_EMBEDDINGS}.{i}.{EMBEDDING_TEXT}
to store data:This causes only the last batch of data to be saved, while data from earlier batches is overwritten and lost.
To Reproduce
llamindex
to construct aVectorStoreIndex
.Expected Behavior
All vector data should be retained, ensuring no data is lost during the embedding process.
Screenshots
(No screenshots provided.)
Desktop (please complete the following information)
Additional Context
The issue occurs due to the handling of keys in the event processing, where unique keys are not assigned to each batch's data.
The text was updated successfully, but these errors were encountered: