/closes #6 /claim #6
A video demonstrating the test process: Link to video
Includes both the worker invoke
and worker stream
To run the tests, follow these steps:
docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
docker exec -it ollama ollama pull moondream # for image test
docker exec -it ollama ollama pull qwen3:1.7b # for text tests
cargo make build
golem-cli app deploy --build-profile ollama-debug
Test 1,2,3,5,6: âś…
Test 4: ⚠️ Partial Success
Test 4 (Tool Use): Ollama’s streaming doesn’t fully support tool calls, Refer to API. The code is there, when Ollama supports this, our code should work as is.
Test 5 (Vision with Moondream): This test uses the moondream
model (only one that will work in my system) to see if image understanding (vision) works via Ollama. moondream
is an older, small model, which works in my system, it doesn’t know Hungarian, but it does reply with some vague idea of the image. (Repeat the test, it gets it somewhat right in 3-4 tries)
Default response for test5 is empty by default, but you can see the reply in the stream. Shouldn’t it output the reply ?
Sending Images to Ollama:
Ollama API
/generate /chat
) and an experimental one that acts like OpenAI’s (/v1
)./v1
one (like OpenAI) for chat, vision (/v1/chat/completions
), as streaming in default api uses ndjson and not sseAjay RV
@Nanashi-lab
Golem Cloud
@golemcloud
Ambati Sahithi
@asr2003