Ollama

Getting started with Ollama.

Ollama is deployed in ExCL as a module. To use Ollama, load the module, and then you have access to the ollama CLI interface.

Load the Ollama module with:

module load ollama

Ollama has a server component which stores files in its home. This server component should be launched using a service account by ExCL admin, since it provides ollama for the entire system. Ollama is already running on some of the workers in ExCL. See the output from the model load for an up-to-date list. Contact excl-help@ornl.gov if you would like ollama to be available on a specific system.

Ollama API

Examples of using the Ollama API can be found at ollama-python/examples/chat.py.

Last updated

Was this helpful?

Revision created

Updated notes.