This add-on for AI Engine adds support for Ollama. Whatâs Ollama? Itâs an LLM you can run on your local machine or server. Itâs incredibly fast, easy to install, and makes you independent from any remote servers and services. Itâs actually the closest you can get to⌠âApple Intelligenceâ! đŹ
How to use it
Once installed, youâll find that a new Type is available in Environments for AI. Pick it, and set the Endpoint. If you installed Ollama locally (same as your WordPress), by default, this endpoint should be âhttp://127.0.0.1:11434â.
When this is done, you should be able to Refresh Models. By default, Ollama should come with one model, llama3. Once it is refresh, you can use this new environment anywhere in AI Engine!
Model Support
It works with all the models, but itâs recommended and well-tested with:
- LLama: Chat, Functions
- Llava: Chat, Image Vision
- Mistral: Chat
- Gemma: Chat
For more information about Ollama, please visit their official website.