Kobold Quickstart Guide

KoboldAi icon

Don't want to read?

Get koboldcppᴬᴹᴰ ᵛᵉʳ⋅ (regular .exe), download this and plug it in. Set your settings to these in Kobold Lite.

Temporary free Command-R+

Cohere offers free trial of their newest 104B model through their API: https://dashboard.cohere.com/ . Short instruction, Kobold Lite supports it too now. This presumably won't last long, so take advantage of it.

Introduction

This is a short complication of various Kobold links — an attempt to streamline and condense all the information in a simple format. The page is intended for people who want to try local/cloud alternatives to NAI but don't have much experience with it. You have a couple of different options, and they are listed from easiest/cheapest to harder and more expensive. The first section is for people completely clueless about local text generation, you can skip it if you know what quantization and VRAM is.

Model size and hardware requirements

Models are evaluated on their parameters size, in billions. The letter B coupled with the number shows how large and compute expensive a model is. The bigger the number, the better the model (usually). 7B/13B are considered small, weaker models, 30B/45B midrage (much better), 70B and above the best and the most demanding. To go above 100B on a local hardware would require you an enormous investment, so don't do that right away. You have to accept that you won't be able to run something on your hardware. The question is, what is enough to satisfy you.

Your biggest bottleneck are hardware requirements. Generally, if you don't have at least 24 GB VRAM, you will be stuck with either a smaller (dumber) model or longer generation times, if you want to use a bigger (smarter) model. You mostly want to run the biggest model you can fit in your VRAM. If you have neither that nor enough normal RAM to load the model, you won't be able to use it at all.

For example, 12 GB VRAM allows you to run 13B models, and if you have less than that, you have to either use 7B or 10.7B. 24 GB VRAM and above is where you can start running 30B, 45B and 70B. The way you can fit those models inside your VRAM even though they are much bigger in their raw state is via the compression process called quantization.

Quantization allows to compress the models to smaller sizes with only small losses. Q8_0 makes the model's size in GB to roughly correspond to its Billions of parameters, with numbers lower than 8 representing further compression. At this quant, there is no perceptible difference between raw and quantized model. Q4_K_M is considered the point past which model's degradation is starting to become worse and much more noticeable. Extreme quantization like Q2_K is advised against, but you can always experiment with a model you really like and see if responses are satisfactory for your needs. Even though Q4_K_M is used the minimum, it doesn't mean that Q3_K_S is unusable, even if it's a little dumb.

With the above two paragraphs in mind, the simplest way to decide on your hardware requirements is to check if Q4_K_M version of the model fits into your VRAM (allow some leeway for context size). If not, you can either go with smaller model or lower quant: the choice is yours.

The best value GPU you can buy is used 3090, with 3060 being the second best for cases of abject poverty. Don't let the others to cloud your judgment with cheaper old cards and AMD 32 GB GPUs — they are not worth it. Buy a used 3090 from Ebay or another marketplace and enjoy.

The current storytuned models are either Mixtral 8x7B - Holodeck (Mistral 7B for low VRAM), BagelMIsteryTour-v2-8x7B (there's also a merge between the two) or Dreamgen's Opus family of models: 7B, 34B and 70B. Holodeck is a completion only model, use Bagel/Donut or Opus if you want instruct capabilities.

But wait, that's so many models, how could you possibly choose just one?
Choose the one with the best sounding name.
You can't. The only way to know which one is the best for you is to try it. It will take you some time, so be ready for a lot of waiting and inconvenience if you hard drive and Internet speed bound.

If you use cloud providers, your selection is greatly limited and customization lacking. You can try Mixtral Instruct, Yi, Claude 3 family and Command-R.

koboldcpp

Generation is done on the CPU, with an option of offloading the model into VRAM. The more of it you offload, the faster it will run. If you can load an entire model into VRAM, you can also try to use GPU-only inference described below for blazing fast prompt processing speeds, but that requires a lot more setup and is generally not recommended. Generation speed for fully offloaded GGUF model and EXL2 model are almost identical.

All you need to do is to download koboldcpp from its releases page (or its AMD fork), download a .gguf quantization of your favorite model that you can run at tolerable speeds (for (You)) and select it via the GUI or drag and drop it.

The repository readme should cover the basics, but if you have any questions, koboldccp has a Github wiki page with more detailed explanations.

Cloud options

Kobold Horde

You can use models generously hosted by other people for free. Because everyone else can request a generation, queue waiting times can be long, and you may only get to use it without delay during certain times of the day.

Keep in mind that even though Horde host can't see what you are generating by default, the software can be modified to do that. Don't send any sensitive information over Horde and always be ready that the host is reading what you are writing even if the chance is fairly small. Since your IP is not sent to the host at all, they can't get any identifying information about you, unless you provide it to them yourself).

To actually use Horde, click the "AI" button in the upper left corner. Choose the model you want and wait for your generation. You are encouraged to register a Horde account following "Register New User" link and do what you can to earn kudos. The details of what kudos are, how to earn them and what is Horde exactly, can be found on their FAQ page
You also can use Horde from either locally hosted Kobold Lite interface or other software that supports it.

Kobold Runpod

You can rent GPUs out for your own needs, and pay for them per hour. Prices vary, your usage is private and you won't get banned for your generations. Only recommended if you know what you are doing and can reign in your AI addiction.

A separate rentry page is fully dedicated to this option.

Vast.ai

The current cheapest GPU rental option. Has more different cards to choose from. Has a KoboldAI template.

koboldcpp Runpod

You can also use Runpod with koboldcpp instead of KoboldAI. All details on setting it up can be found in the readme (just click the button after following the link).

OpenRouter

New model provider option, supposedly the cheapest and the easiest to use. Mostly hosts chat and instruct models. Unlike Runpod and other GPU rentals, you pay for model tokens, not GPU time. Is supported by Kobold Lite. Simply choose it after clicking "Use Custom Endpoint" in "AI" tab.

Together AI

Has a $25 trial.

Their API is supported in Kobold Lite if you use OpenAI endpoint. Simply put https://api.together.xyz/v1 instead of OpenAI URL and paste in your Together key. Uncheck "Use ChatCompletions API" so you can use it raw.

koboldcpp on Google Colab

Will run .gguf quants of the models using Google's hardware. You can load other models (including NSFW), as long as you correctly provide the direct link following the default model example. Keep it to 13B Q4_K_M sizes unless you know what you are doing. It should support any model current main branch koboldcpp does. Simply wait for the cell to execute and follow the remote tunnel link that is printed once it's done.

Don't expect any privacy, everything is sent straight to Google and will certainly be used to train their own models. Don't share anything you wouldn't send over Horde or any private information (Google already knows who you are). The risk of getting your Google account banned is very low, but not zero. You can never know when a corporation decides to flip the switch and bring down the ban hammer.

There is also older KoboldAI colab GPU version that can load any Hugging Face model, as long as you give it a valid link. Only use it if you have a specific reason for it.

Other cloud providers

There are other cloud hosts you can use with Kobold Lite. That includes everything under "Custom AI Endpoint". You can pay for them or get them for free. Or you can do both. A python script that provides OpenAI and KoboldAI compatible endpoints for Claude (both the legit API and proxies) is also available.

Use the above only at your own risk.

KoboldAI

GPU-focused local client, the very first one.
Stuck in a perpetual in-development stage and is missing a lot of newest features for loading models. UI2 is still in beta and has a lot of bugs.

With this client, you are to use exl2 quants you can get from Hugging Face. You need to download the whole folder from the repo and put it as the subfolder in KoboldAI/models. It's very important to include "tokenizer.json" file. For example, if you have 24GB of VRAM, you might want to get 3.5 bpw quants if you want to run a Mixtral finetune (if you want to use another model, simply type "MODEL NAME exl2 into HuggingFace search bar"). Running GPU inference makes your prompt processing blazingly fast, but you won't get any significant gains on generation speed itself. Therefore, if you don't edit your context a lot, .gguf models and CPU inference remains the recommended option.

Text generation web UI

An alternative to KoboldAI. Has more active development and supports a lot more modern features. You also need to get exl2 quants for it (although it supports all other options, including CPU generation). The interface is lacking, but it can be used with either Mikupad or Sillly Tavern.

Some of the above options are not recommended by certain users who use AI for storywriting with prose. Try them and see if they satisfy your needs.

Edit
Pub: 23 Feb 2024 04:40 UTC
Edit: 07 Apr 2024 13:09 UTC
Views: 2388