LlamaIndex Integration
EngramVectorStore from engram.sdk.llama_index implements BasePydanticVectorStore.
Install
bash
pip install llama-index-core engram-subnet
Basic usage
python
from llama_index.core import VectorStoreIndex, Documentfrom llama_index.core.storage.storage_context import StorageContextfrom engram.sdk.llama_index import EngramVectorStorestore = EngramVectorStore(miner_url="http://127.0.0.1:8091")storage_context = StorageContext.from_defaults(vector_store=store)documents = [Document(text="Bittensor is a decentralised ML network."),Document(text="TAO tokens reward miners and validators."),Document(text="Engram stores embeddings on Bittensor."),]index = VectorStoreIndex.from_documents(documents,storage_context=storage_context,)
Query engine
python
query_engine = index.as_query_engine()response = query_engine.query("How does Bittensor distribute rewards?")print(response)# Retrieve without generatingretriever = index.as_retriever(similarity_top_k=5)nodes = retriever.retrieve("TAO tokenomics")
Note
The
delete() method is a no-op in Engram — vectors are content-addressed and immutable once stored.engram docs · v0.1edit on github →