Skip to main content
🦾 OpenLLM lets developers run any open-source LLMs as OpenAI-compatible API endpoints with a single command.
  • πŸ”¬ Build for fast and production usages
  • πŸš‚ Support llama3, qwen2, gemma, etc, and many quantized versions full list
  • ⛓️ OpenAI-compatible API
  • πŸ’¬Β Built-in ChatGPT like UI
  • πŸ”₯ Accelerated LLM decoding with state-of-the-art inference backends
  • πŸŒ₯️ Ready for enterprise-grade cloud deployment (Kubernetes, Docker and BentoCloud)

Installation

Install openllm through PyPI
pip install -qU  openllm

Launch OpenLLM server locally

To start an LLM server, use openllm hello command:
openllm hello

Wrapper

from langchain_community.llms import OpenLLM

server_url = "http://localhost:3000"  # Replace with remote host if you are running on a remote server
llm = OpenLLM(base_url=server_url, api_key="na")
llm("To build a LLM from scratch, the following are the steps:")

Connect these docs programmatically to Claude, VSCode, and more via MCP for real-time answers.