Had a little epiphany around using large language models locally; how the number of parameters relates to amount of GPU VRAM. The relationship is indirect and exponential so dial-twiddling is fussy. Upshot is larger parameter models working reasonably well on a laptop with an NVIDIA GPU. No smoke.
#LLM