Migrating to llama.cpp drop ollama and run models directly
Running flux locally on a mac lets get those gpus running