Question for those of you who host a LLM by themselfs with Ollama, llama.cpp and use it for example for generating alt texts for images.
What LLM do you recommend? Which one generates a good description for screen reader users with the least amount of computing?
Whats your experience with that? Bonus points for LLM's which perform really good in CPU only situations.