Works like a charm on a Fedora + podman compose !! I had to tweak the ollama-service a tiny bit with Philippe Charrière's help:
services:
# https://cheshirecat.ai/local-models-with-ollama/
ollama-service:
container_name: ollama_pi_local
user: root
image: ollama/ollama:0.1.27
volumes:
- ollama-data:/root/.ollama
ports:
- 11434:11434
volumes:
ollama-data: