A local LLM running on x86 or M2 would be great. But API callouts while waiting for the silicon/Cuda migrations are fine too.
A local LLM running on x86 or M2 would be great. But API callouts while waiting for the silicon/Cuda migrations are fine too.