

2·
7 days agoMy go to for messing with chat bots is Kobold that’ll let you split the work between multiple GPUs. I get the impression the actual processing is only done on one but it lets you load larger models with the extra memory.


My go to for messing with chat bots is Kobold that’ll let you split the work between multiple GPUs. I get the impression the actual processing is only done on one but it lets you load larger models with the extra memory.


I was using a Nvidia 3060 for a while, then had 2 in one box, then switched to a 3090.
The amount of vram is a big factor for decent performance. Getting it to not sound like a predictably repetitive bot though is a whole separate thing that is still kind of elusive.

Because if you said jetstream destabilization and poler vortex he would try and claim they where made up words by some shadowy conspiracy in order to play to his education hating base?
You get 50%, you get 75%, you get 225%…
Guy pulls numbers out of his ass for whatever dramatic effect he thinks he can make and then blames others as why the economy is in turmoil.