Ollama
Getting started with Ollama.
Last updated
Getting started with Ollama.
Last updated
Ollama is deployed in ExCL as a module. To use ollama, load the module, and then you have access to the ollama
cli interface.
Load the Ollama module with:
Ollama has a server component which stores files in its home. This server component should be launched using a service account by ExCL admin, since it provides ollama for the entire system. Ollama is already running on some of the workers in ExCL. See the output from the model load for an up-to-date list. Contact excl-help@ornl.gov if you would like ollama to be available on a specific system.
When interacting with the Ollama server via the REST API in ExCL, you need to unset the http_proxy
and https_proxy
environment variables, since you are trying to connect to an internal http server instead of a remote one.
An examples of using the Ollama API can be found at ollama/examples/python-simplegenerate at main · ollama/ollama.