Llama.cpp
llama.cpp python library is a simple Python bindings for
@ggerganovllama.cpp.This package provides:
- Low-level access to C API via ctypes interface.
 - High-level Python API for text completion
 
OpenAI-like APILangChaincompatibilityLlamaIndexcompatibility- OpenAI compatible web server
 
- Local Copilot replacement
 - Function Calling support
 - Vision API support
 - Multiple Models
 
%pip install --upgrade --quiet  llama-cpp-python
from langchain_community.embeddings import LlamaCppEmbeddings
API Reference:LlamaCppEmbeddings
llama = LlamaCppEmbeddings(model_path="/path/to/model/ggml-model-q4_0.bin")
text = "This is a test document."
query_result = llama.embed_query(text)
doc_result = llama.embed_documents([text])
Related
- Embedding model conceptual guide
 - Embedding model how-to guides