Skip to content

Support native macOS (Apple Silicon) execution without Docker #608

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
davidfowl opened this issue Mar 30, 2025 · 1 comment
Open

Support native macOS (Apple Silicon) execution without Docker #608

davidfowl opened this issue Mar 30, 2025 · 1 comment
Labels
awaiting response Waiting for the author of the issue to provide more information or answer a question

Comments

@davidfowl
Copy link
Contributor

Right now, running Ollama via Docker on macOS doesn’t use the GPU because Docker can’t access Apple Silicon GPUs. That’s a pretty big performance hit, especially when working with larger models.

This blog post lays it out clearly: https://chariotsolutions.com/blog/post/apple-silicon-gpus-docker-and-ollama-pick-two/

When you run Ollama natively, it detects and uses the Metal backend:

level=INFO msg="Dynamic LLM libraries [metal]"

But in Docker:

level=INFO msg="no GPU detected"

It’d be great if Ollama could run natively on macOS without Docker, so we can take advantage of Apple’s GPU acceleration. Performance and dev experience would be much better.

@aaronpowell
Copy link
Member

Would you expect it to download/install Ollama if it's not already installed?

Can you specify the port of the Ollama host via the CLI? That'd be important for ensuring you don't have port conflicts on startup.

A current workaround would be to use the ConnectionStringResource and provide the endpoint for Ollama run from outside of Aspire.

@aaronpowell aaronpowell added the awaiting response Waiting for the author of the issue to provide more information or answer a question label Apr 8, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
awaiting response Waiting for the author of the issue to provide more information or answer a question
Projects
None yet
Development

No branches or pull requests

2 participants