A40
.
11434
to the list of exposed ports. This port is used by Ollama for HTTP API requests.
OLLAMA_HOST
0.0.0.0
ollama serve
part starts the Ollama server, making it ready to serve AI models.
Now that your Ollama server is running on your Pod, add a model.
ollama run
command:
[model name]
with the name of the AI model you wish to deploy. For a complete list of models, see the Ollama Library.
This command pulls the model and runs it, making it accessible for inference. You can begin interacting with the model directly from your web terminal.
Optionally, you can set up an HTTP API request to interact with Ollama. This is covered in the next step.
[your-pod-id]
with your actual Pod Id.[your-pod-id]
with your actual Pod Id.
Because port 11434
is exposed, you can make requests to your Pod using the curl
command.
For more information on constructing HTTP requests and other operations you can perform with the Ollama API, consult the Ollama API documentation.