Local AI Chatbot experience + resources
Here I collected some of the things that worked for me and were relatively lightweight.
I wouldn't say the model behaves as good as GPT-4 or even GPT-3.5, but it is fine for short and shallow conversations, considering that it runs right on the laptop, no need for internet connection.
My hardware: GPD Pocket 3 (16Gb RAM, i7-1195G7). I run models on the CPU, because any respectful GPU is absent.
Chat UI: Amica
Server: llama.cpp
Model: Meta AI's LLaMA v2.
Although 13B parameters run just fine for me in the console (2-3 tokens per second), with the chat UI open the performance becomes abysmal (i.e. 1 token per 5-10 seconds or so).
So I opt out for a 7B parameter configuration. As per my subjective judgement, it is decent enough. Right on the edge, so to speak.
I have also tried 3B parameter configuration, but it is not good enough for holding at least a plausible chat. The only good thing is it can run on a Raspberry PI (32-bit raspbian) :)
Upd. 13B performance is somehow better in web chat UI. The built-in one, not Amica.