Prerequisites
Before proceeding, ensure the following prerequisites are met:- Install MindsDB locally via Docker or Docker Desktop.
- To use Ollama within MindsDB, install the required dependencies following this instruction.
- Follow this instruction to download Ollama and run models locally.
Here are the recommended system specifications:
- A working Ollama installation, as in point 3.
- For 7B models, at least 8GB RAM is recommended.
- For 13B models, at least 16GB RAM is recommended.
- For 70B models, at least 64GB RAM is recommended.
Setup
Create an AI engine from the Ollama handler.ollama_engine
as an engine.
If you run Ollama and MindsDB in separate Docker containers, use the
localhost
value of the container. For example, ollama_serve_url = 'http://host.docker.internal:11434'
.Usage
The following usage examples utilizeollama_engine
to create a model with the CREATE MODEL
statement.
Deploy and use the llama3
model.
First, download Ollama and run the model locally by executing ollama pull llama3
.
Now deploy this model within MindsDB.
Models can be run in either the ‘generate’ or ‘embedding’ modes. The ‘generate’ mode is used for text generation, while the ‘embedding’ mode is used to generate embeddings for text.However, these modes can only be used with models that support them. For example, the
moondream
model supports both modes.By default, if the mode is not specified, the model will run in ‘generate’ mode if multiple modes are supported. If only one mode is supported, the model will run in that mode.To specify the mode, use the mode
parameter in the CREATE MODEL
statement. For example, mode = 'embedding'
.Next StepsGo to the Use Cases section to see more examples.