ollama / start.sh
Germano Cavalcante
Initial commit
a8144f6
raw
history blame
570 Bytes
#!/bin/bash
# Set environment variables for the ollama server
export OLLAMA_HOST=0.0.0.0
export OLLAMA_ORIGINS=https://projects.blender.org
# Start the Ollama service in the background
ollama serve &
# Wait for the service to initialize
sleep 10
# Download the required file
curl -fsSL https://huggingface.co/lmstudio-community/Llama-3-Groq-8B-Tool-Use-GGUF/resolve/main/Llama-3-Groq-8B-Tool-Use-Q4_K_M.gguf?download=true -o llama.gguf
# Create the model using Ollama
ollama create llama3.1 -f Modelfile
# Keep the container running indefinitely
tail -f /dev/null