I believe it's time to retire my #selfhosted #Ollama instance. My #RTX3080 can handle some models, but not the models (and sizes) that I actually need/want to use. Perhaps I can slap that 3080 into another rig for gaming or something...or maybe I can try using it for ComfyUI or image generation related. 