Question for those of you who host a LLM by themselfs with Ollama, llama.cpp and use it for example for generating alt texts for images.
What LLM do you recommend? Which one generates a good description for screen reader users with the least amount of computing?

Whats your experience with that? Bonus points for LLM's which perform really good in CPU only situations.

#selfhosting @homelab @homelab_de #llm #accessibility