Using llama-cpp-python server with LangChain

Very quick and short one. I was trying to make something with LangChain. I already had a server running LLaMA 2 Chat using llama-cpp-python. It happens to provide a OpenAI like API. There must be a way to abuse the code to make it work with LangChain. Yet half and hour of Googling turned up nothing. So I decided to read the code.

And... it's easy. The openai python package reads a OPENAI_API_BASE environment variable to know where to send the requests. So, just set it to the LLaMA server and you're good to go. Also since llama-cpp-python doesn't to authentication, you can just leave the OPENAI_API_KEY to be antyhing. As long as it makes the Python package happy.

export OPENAI_API_BASE="http://llm.your.domain.com/v1"

=> Using llama-cpp-python server with LangChain

That's all. Hope this helped someone searching and can't find anything.

Proxy Information
Original URL
gemini://gemini.clehaxze.tw/gemlog/2023/09-25-using-llama-cpp-python-server-with-langchain.gmi
Status Code
Success (20)
Meta
text/gemini
Capsule Response Time
1519.929847 milliseconds
Gemini-to-HTML Time
0.561993 milliseconds

This content has been proxied by September (3851b).