So a rather common question we get is “Can you run the mod offline without needing to pay to use an API?”. The answer is Yes, but…..
Running an offline LLM is possible thanks to a neat program called Koboldcpp. It can be appealing being able to run the mod on your own hardware and also if you are concerned about data privacy. However it can cause you a lot of headache if you do not know what you are doing.
It has a lot of drawbacks compared to using an online model:
UNLESS YOU REALLY KNOW WHAT YOU ARE DOING.
WE DO NOT RECOMMEND USING KOBOLDCPP WITH CHIM.
Please just use OpenRouter if it's your first time setting up the mod. It is much much easier to get working.
We have kept it in as a feature because we have a lot of advanced users, who do know what they are doing, who enjoy tinkering with it. But if it's your first time setting up this mod, just use OpenRouter. You will save yourself so much trouble trying to configure an offline LLM compared to the plug and play capability of an online service. For $5 you will get a lot of playtime out of it.
However if you are still determined to use Kobold CPP, then below we have a short guide created by hey_danielx on what you need to do to set up a working configuration.
Users have also noticed that SSE DisplayTweaks helps Kobolod run faster while playing Skyrim.
Use this model: https://huggingface.co/bartowski/magnum-12b-v2-GGUF
(Q5_K_L quant is a good starting point)
USE THESE SETTINGS FOR https://huggingface.co/bartowski/magnum-12b-v2-GGUF