Installation
LLM-IE Python package and web application installation guide.
Python package
The Python package is available on PyPI.
Note that this package does not check LLM inference engine installation nor install them. At least one LLM inference engine is required. There are built-in supports for LiteLLM, Llama-cpp-python, Ollama, Huggingface_hub, OpenAI API, and vLLM. For installation guides, please refer to those projects. Other inference engines can be configured through theInferenceEngine
abstract class. See LLM Inference Engine section.
Web Application
Docker
The easiest way to install our web application is 🐳Docker. The image is available on Docker Hub. Use the command below to pull and run locally:
docker pull daviden1013/llm-ie-web-app:latest
docker run -p 5000:5000 daviden1013/llm-ie-web-app:latest
Open your web browser and navigate to: http://localhost:5000
If port 5000 is already in use on your machine, you can map it to a different local port. For example, to map it to local port 8080:
Then visit http://localhost:8080
Install from source
Alternatively, pull the repo and build the required environment locally.