Sukino's Findings: A Practical Index to AI Roleplay
Finding learning resources for AI roleplaying can be tricky, as most of them are hidden away in Reddit threads, Neocities pages, Discord chats, and Rentry notes. It has a lovely Web 1.0, pre-social media vibe to it, with nothing really indexed or centralized and always something cool buried somewhere you haven't discovered yet.
To make things a little easier, I've compiled a list of interesting, up-to-date information about it. Think of it as a crash course to help you get a modern AI roleplaying setup, understand how everything works and where to find things. Want to know more? Check out my Guides page, where I share little quality of life things I have discovered.
If you have any feedback, want to talk, make a request, or share something, reach me at: sukinocreates@proton.me or @sukinocreates
on Discord.
Latest Updates:
2025-03-22 — I made conversions of Deepseek presets for Text Completion connections, check them out. Added warning that Gemini got worse at roleplaying lately.
2025-03-20 — Addeddebased-AI
andtheatreJB
presets. addedWyvernChat
to character card providers.
2025-03-18 — TheLocal LLM/Models
section has been reworked with a new tool, but the information itself is the same.
2025-03-15 — Changed 24B recommendations to Dan's Personality Engine and Cydonia 2.1. Added more 12B recommendations for variety, Rocinante 1.1 and NemoMix Unleashed.
2025-03-13 — Updated list of Presets. Expanded theLocal LLM/Models
section. AddedGuided Generations
.
Getting Started
Picking an Interface
First, you will need a frontend, the interface where the roleplaying takes place and where your characters live.
I will only recommend solutions that are open source, private, secure, well maintained, and don't lock you into a closed ecosystem. So if you've heard of a service that's not listed here, it's probably because it doesn't meet these criteria.
- Install SillyTavern: https://docs.sillytavern.app/installation/ — SillyTavern stands as the de facto frontend for AI roleplaying. While alternatives exist, it remains the most mature and feature-rich platform, consistently receiving updates first and offering extensive customization options with broad system compatibility and robust community support. Follow the official guide for step-by-step instructions on Windows, Linux, Mac, Android, or Docker. For iOS users, see the solution below.
- Access SillyTavern Remotely Via Tailscale: https://sillytavernai.com/tailscale-config / https://docs.sillytavern.app/administration/tunneling — Tailscale is a private and secure tunnel that connects all your devices like a LAN connection, but over the Internet. This means that you can host SillyTavern on one device, and access the same instance on all your other devices anywhere you have an Internet connection, and maybe even share it with your friends. This is the best way to use it on Android, and the only way to really get it on iOS, if you have a dedicated device to host it — like a computer, unused phone, Raspberry Pi, or homeserver. You can even rent a small, inexpensive remote server or VPS to host it on, if you are tech savvy enough, it's pretty lightweight.
- Or Use Agnastic: https://agnai.chat/ — If you really can't install SillyTavern, or you just want a simple online solution, Agnastic is becoming a good alternative. It's free, can be used without an account and runs completely in your browser. It even has some free models for you to get your feet wet — there are better free models in the next section, so don't pick it just because of this.
- or RisuAI: https://risuai.xyz/ — Another online alternative. Has a different set of features than Agnastic, and some users find the UI more friendly, so it might be more to your liking.
There's nothing stopping you from starting with these online frontends and later migrating to SillyTavern if you feel the need for a more complete solution. Just keep in mind that you'll miss out on most of the modern and advanced features, and that most of the content and setups you find online won't apply to you.
Throughout this guide, I'll assume you're using SillyTavern, but the instructions should be easily applicable to the alternatives—you'll just need to look for the equivalent options.
Setting Up an AI Model
If You Want to Run an AI Locally
It's uncensored, free, and private. Requires a computer or server with a dedicated GPU or a Mac with an M-series chip. If you don't know if you have dedicated GPU, Google or ask ChatGPT for instructions on checking for your system.
You'll need a backend, the program that will run your AI models and connect to your frontend via a local API. There are two main model formats to pick from, GGUF and EXL2. If you don't have a preference yet, go with GGUFs, they are easier to find, easier to use, and have more sizes to fit all memory sizes.
Choose a backend and go pick a model and a suitable preset.
- KoboldCPP: https://github.com/LostRuins/koboldcpp — Runs GGUF models. Don't know what to pick? Go with this one. Designed with roleplaying in mind, so it has some exclusive features for us roleplayers that will come up later in the guide. Comes with its own roleplaying frontend that you can use if you want to, but you don't have to interact with it. Read the notes on the release page to know which version you need to download.
- I have a guide that will help you set it up and optimize it to your system, check it out.
- TabbyAPI: https://github.com/theroyallab/tabbyAPI — Runs EXL2 models. Probably will be the most performant if you have enough VRAM to run everything smoothly.
- LM Studio: https://lmstudio.ai/ — Runs GGUF models. Pretty barebones, but has it's fans for how easy it is to use, and for being able to download and manage the models within it's UI.
- TextGen WebUI/Oobabooga: https://github.com/oobabooga/text-generation-webui — Runs GGUF and EXL2 models. The most versatile and it's strength is having the best integrated UI to chat with the AI model.
If You Want to Use an Online AI
This is where censorship and privacy become an issue, as you will be sending everything to these services, and they can log your activity, block your requests, or ban you at will. Stay safe, use burner accounts if you feel like it would be bad to have your sessions tied to your name, and be careful not to accidentally send sensitive information, as most of the time your data will be used to train new AI models.
You'll need a service that provides the AI model of your choice and an API key to connect to it with your frontend.
Choose a service and go pick a suitable preset.
- The Free Ones: These change all the time, but I will try to keep this updated with the options I know of.
- Google Gemini: https://aistudio.google.com/apikey — Google currently offers free API keys through AI Studio. Gemini used to be the best roleplaying models available for free, but Google has gradually made them worse. Has many security checks, so a good preset is essential, and you may still encounter refusals. Use the model version suggested by your chosen preset, as the models are updated frequently and the ideal one for roleplaying changes. Requires a Google account, and your data will be used for training (if used outside the UK/CH/EEA/EU), but since it's Google, you can't expect much else.
- MistralAI: https://console.mistral.ai/api-keys — Mistral currently offers trial API keys through Le Plateforme. Mistral Large 2411 is their best model. Requires opting into data training and may ask for phone number verification.
- Cohere: https://dashboard.cohere.com/api-keys — Cohere currently offers evaluation API keys. Command R+ 104B (not 08-2024) is their best model. Requires registration and is rate limited, read their documentation to know more.
- Model Providers With Free Key Rotations: https://github.com/cheahjs/free-llm-api-resources — There are providers/revendors that host the same AI models that people run locally, and offer some of them for free in a rotation. However, you cannot verify the real quality of the models; they may be serving a very low-quality version to free users.
- KoboldAI Colab: https://koboldai.org/colabcpp — You can borrow a GPU for a few hours to run KoboldCPP at Google Colab. It's easier than it sounds, just fill in the fields with the desired GGUF model link and context size, and run. They are usually good enough to handle small models, from 8B to 12B, and sometimes even 24B if you're lucky. Check the section on where to find local models to get an idea of what are the good models.
- AI Horde: https://stablehorde.net/ — A crowdsourced solution that allows users to host models on their systems for anyone to use. The selection of models depends on what people are hosting at the time. It's free, but there are queues, and people hosting models get priority. By default, the host can't see your prompts, but the client is open source, so they could theoretically modify it to see and store them, though no identifying information (like your ID or IP) would be available to tie them back to you. Read more on their FAQ to be aware of any real risks.
- The Paid Ones: Most of these options operate on a pay-per-request model, so the more you play, the more expensive it gets.
- There are providers/revendors that host the same AI models that people run locally, at all price points. The most famous is OpenRouter, but you can find alternatives if you look around, including cheaper and subscription based ones. Shop around and check the section on where to find local models to get an idea of what are the good models.
- You can also pay for the big, corporate models like GPT, Claude and Deepseek, they are pretty smart and will give you the best experience you can get. But only go this route if you have disposable income because they are quite expensive and it turns into a money sink pretty fast. And remember, you need the API key, so don't buy a ChatGPT subscription or anything like that.
- /aicg/ meta: https://rentry.org/aicg_meta — Comparison of how the different services/models perform in roleplay. Don't take this as gospel, they vary depending on the preset and bots you use, but it can help you set your expectations for what you can pay for.
Your model's provider/proxy isn't available via Chat Completion in your frontend?
You'll need to find out if they offer an OpenAI-compatible endpoint. Basically, it mimics the way OpenAI's ChatGPT connects, adding compatibility with almost any program that supports GPT itself. Check their documentation looking for anendpoint address
, it should look something like thishttps://api.provider.ai/v1
. If they have one, selectCustom (OpenAI-compatible)
as your chat completion provider, and manually enter that address and your API key. If the model list loads, you are golden, just select the right model there.
Where to Find Stuff
Chatbots/Character Cards
Chatbots
, or simply bots
, come in image files, or rarely in json files, called character cards
. The chatbot's definitions are embedded in the image's metadata, so never convert it to another format or resize it, or it will become a simple image. You simply import the character card into your roleplaying frontend and the bot will be configured automatically.
- Chub AI: https://chub.ai/ — This is the primary hub for chatbot sharing, but it's overwhelmed with frustratingly low-quality bots. It's hard to find the good stuff without knowing who the good creators are. So, for a better experience, create an account and follow creators whose bots you enjoy.
- Chub Deslopfier: https://gist.github.com/khanonnie/b357f20bfe4e920d8e05fd47f1e6fa75 — Browser script that tries to detect and hide extremely low quality cards.
- Chatbots Webring: https://chatbots.neocities.org/ — A webring in 2025? Cool! Automated index of bots from multiple creators directly from their personal pages. Could be a great way to find interesting characters without drowning in pages of low-effort sexbots on Chub. I mean, if the creator went to the trouble of setting up a website to host their bots, they must be into something, right?
- Anchored Bots: https://partyintheanchorhold.neocities.org/ — Consistently updated list of bots shared on 4chan without having to access 4chan at all, what a blessing.
- The meta list of various bot lists from different boards: https://rentry.org/meta_bot_list — More 4chan bots.
- WyvernChat: https://app.wyvern.chat/ — A strictly moderated bot repository that is gaining popularity.
- Character Tavern: https://character-tavern.com/ — Community-driven platform dedicated to creating and sharing AI Roleplay Character Cards.
- AI Character Cards: https://aicharactercards.com/ — Promises higher-quality cards though stricter moderation.
- RisuRealm Standalone: https://realm.risuai.net/ — Bots shared through the RisuRealm from RisuAI.
- JannyAI: https://jannyai.com/ — Archive of bots ripped from JanitorAI. If you are a migrating user, this may be of interest to you.
- PygmalionAI: https://pygmalion.chat/explore — Pygmalion isn't as big on the scene anymore, but they still host bots.
- Character Archive: https://char-archive.evulid.cc/ — Archived and mirrored cards from many various sources. Can't find a bot you had or that was deleted? Look here.
- Chatlog Scraper: https://chatlogs.neocities.org/ — Want to read random people's funny/cool interactions with their bots? This site tries to scrape and catalog them.
Local LLM/Models
Figuring Out Which Models You Can Run
Want to run a model locally, but are confused by all those names and numbers? No worries! Here's a quick crash course, plus two tools that will help you find the perfect model. First, you just need to understand these four key concepts:
Total VRAM
is the memory you have available in GPU, your graphics card. This is different than your RAM memory. If you don't know how much memory you have, or if you have dedicated GPU, Google or ask ChatGPT for instructions on checking for your system.- In roleplay, the
Context Length
is how many past messages the AI can hold in memory, measured intokens
, between a syllable and a word.8192
tokens is pretty good; users generally prefer16384
for long roleplaying sessions, but you may need to choose a worse model to be able to fit everything in your GPU. An oversized context is useless if your model can't use all the information, so don't go beyond 16K for now, as most models compatible with common domestic hardware can't use it effectively. - Models have sizes, calculated in billions of parameters, represented by a number followed by
B
. Largermodel sizes
are generally smarter, but not necessarily better at roleplaying, and require more memory to run. So, as a rule of thumb, a model with 12B parameters is smarter than one with 8B parameters. - Models are shared in various quantizations, or
quants
. The lower the number, the dumber the model gets, but less memory you need to run it. The best balance between compatibility and intelligence for AI roleplaying purposes is a GGUF IQ4_XS (or Q4_K_S if there isn't one available), or an EXL2 between 4.0~4.5 bpw.
Simple, right? Total VRAM, context length, model sizes, and quants. Now we will use this information with one of these two calculators:
- https://sillytavernai.com/llm-model-vram-calculator/ — This tool is the easiest to use. Just enter your Total VRAM and desired Context Size, then click Load Models to see a list of compatible options. Once it loads, sort by Total VRAM and find the highest number followed by B—this indicates the largest model your hardware can run smoothly at IQ4_XS or Q4_K_S. For example, if your system can handle an 8B model, you can run basically any model in that size range or smaller. But I suggest that you choose a
Default Recommendation
bellow instead of the ones suggested by the calculator, their algorithm favors older models not fine-tuned for roleplaying, as they are more widely used and have had more time to gather more reviews and downloads. - https://smcleod.net/vram-estimator/ — If you are a bit more tech-savvy, this calculator is pretty self-explanatory and will let you find the perfect model size and quant for your system. Just adjust the values until the
FP16 K/V Cache
bar fits into the available VRAM of your GPU.
Default Recommendations
These are the most commonly recommended models by 2025-03. They're not necessarily the freshest or my favorites, and there's no one best model for everyone, but they're tried and true. It's a good idea to test and keep a few models around for variety, as small local models can get repetitive over time and different models tend to have different flavors. Choose a model and go pick a suitable preset.
- 7B: SanjiWatsuki/Silicon-Maid-7B (Alpaca Instruct) — GGUF / EXL2
- 7B: SanjiWatsuki/Kunoichi-7B (Alpaca Instruct) — GGUF / EXL2
- 8B: Sao10K/L3-8B-Lunaris-v1 (Llama 3 Instruct) — GGUF / EXL2
- 8B: Sao10K/L3-8B-Stheno-v3.2 (Llama 3 Instruct) — GGUF / EXL2
- 12B: inflatebot/MN-12B-Mag-Mell-R1 (ChatML Instruct) — GGUF / EXL2
- 12B: MarinaraSpaghetti/NemoMix-Unleashed-12B (Mistral V3 Instruct) — GGUF / EXL2
- 12B: TheDrummer/Rocinante-12B-v1.1 (ChatML Instruct) — GGUF / EXL2
- 22B: TheDrummer/Cydonia-22B-v1.2 (Metharme Instruct) — GGUF / EXL2
- 24B: PocketDoc/Dans-PersonalityEngine-V1.2.0-24b (ChatML Instruct) — GGUF / EXL2
- 24B: TheDrummer/Cydonia-24B-v2.1 (Mistral V7 Instruct) — GGUF / EXL2
- Sukino's Banned Tokens: https://huggingface.co/Sukino/SillyTavern-Settings-and-Presets/blob/main/Banned%20Tokens.txt — Hey, it's me! This isn't really a model, but a list of clichés and repetitive phrases that AIs like to use, which you can use to ban them from any model's vocabulary if you've chosen KoboldCPP as your backend. Play a bit without it, then try it, it might improve the responses and make you feel like you've upgraded your model, and it's easy to undo if you don't like the effect. Using this with other backends will mess up your AI responses instead. The list is still being updated, so check back from time to time.
Finding More Models
- HuggingFace: https://huggingface.co/models — This is where you actually download models from, but browsing through it is not very helpful if you don't know what to look for.
- Bartowski/mradermacher: https://huggingface.co/bartowski / https://huggingface.co/mradermacher — I don't know how they do it, but these two keep releasing GGUF quants of every slightly noteworthy model that comes out really quickly. Even if you don't use GGUF models, it's worth checking their profile to see what new models are released.
- SillyTavernAI Subreddit: https://www.reddit.com/r/SillyTavernAI/ — Want to find what models people are using lately? Do not start a new thread asking for them. Check the weekly
Best Models/API Discussion
, including the last few weeks, to see what people are testing and recommending. If you want to ask for a suggestion in the thread, say how much VRAM and RAM you have available, or the provider you want to use, and what your expectations are. - HobbyAnon: https://venus.chub.ai/users/hobbyanon — This page offers a curated list of models of multiple sizes and instruct templates, along with an easy-to-follow tutorial for getting started with KoboldCPP.
Presets, Prompts and Jailbreaks
Always use a good preset that is appropriate for your model of choice. They are also called prompts
or jailbreaks
, although this name can be a bit misleading as they are not just for making these AI models write smut and violence — the NSFW part is usually optional.
LLM models are first and foremost corporate-made assistants, so giving them well-structured instructions on how to roleplay and what the user generally expects from a roleplaying session is really beneficial to your experience. Each preset will play a little differently, based on the creator's preferences and the quirks they found with the models, so try different ones to see which one is more to your liking.
Presets are listed by the model or instruct template with which they are compatible. If you're using a finetune and the instruct template isn't obvious from the model name alone, you can usually find that information on the model's original creator page.
Presets for Text Completion Models
To import these presets on SillyTavern, click on the AI Response Formatting
button, the third one with an A
in the top bar, and press the Master Import
button on the top-right of the window. Make sure the ones you downloaded are selected in the drop-down menus. Always read their descriptions to make sure you don't need to tweak any other setting.
- sphiratrioth666: https://huggingface.co/sphiratrioth666/SillyTavern-Presets-Sphiratrioth — Alpaca, ChatML, Llama, Metharme/Pygmalion, Mistral
- MarinaraSpaghetti: https://huggingface.co/MarinaraSpaghetti/SillyTavern-Settings — ChatML, Mistral
- Virt-io: https://huggingface.co/Virt-io/SillyTavern-Presets — Alpaca, ChatML, Command R, Llama, Mistral
- debased-ai: https://huggingface.co/debased-ai/SillyTavern-settings — Gemma, Llama
- Sukino: https://huggingface.co/Sukino/SillyTavern-Settings-and-Presets — ChatML, Deepseek, Gemma, Llama, Metharme/Pygmalion, Mistral
- The Inception: https://huggingface.co/Konnect1221/The-Inception-Presets-Methception-LLamaception-Qwenception — Llama, Metharme/Pygmalion, Qwen — This one is pretty big, so I wouldn't recommend for small models. Make sure your model is smart enough to handle it.
- CommandRP: https://rentry.org/4y1je_commandrp — Command R/R+
Presets for Chat Completion Models
To import these presets on SillyTavern, click on the AI Response Configuration
button, the first one with the sliders in the top bar, and a windows titled Chat Completion Presets
should pop up — If it has another name, you aren't connected via Chat Completion, fix it first. Now, just press the Import preset
icon on the top-right of the window, and make sure the one you downloaded is selected in the drop-down menu. Always read their descriptions to make sure you don't need to tweak any other setting.
- pixi: https://pixibots.neocities.org — Claude, Deepseek, Gemini
- momoura: https://momoura.neocities.org/ — Claude, Deepseek, Mistral Large (Outdated)
- AvaniJB https://rentry.org/avaniJB — GPT, Gemini
- MarinaraSpaghetti: https://rentry.org/marinaraspaghetti — Gemini
- MarinaraClaude: https://rentry.org/marinaraclaude — Claude
- SmileyJB: https://rentry.org/SmileyJB — Claude, GPT
- Pitanon: https://rentry.org/pitanonprompts — Claude, Deepseek, GPT
- XMLK/CharacterProvider: https://rentry.org/CharacterProvider — Claude, GPT
- Holy Edict: https://rentry.org/Writing_Style — Claude, GPT, Gemini
- Lumen: https://illuminaryidiot.neocities.org/presets — Claude, GPT
- Fluff: https://rentry.org/fluffpreset — Gemini
- DeepFluff: https://rentry.org/DeepFluff — Deepseek
- ArfyJB: https://rentry.org/ArfyJB — Claude, Deepseek, GPT
- CherryBox: https://rentry.org/CherryBox — Deepseek
- Quick Rundown on Large REVISED: https://rentry.org/large-qr-revised — Mistral Large
- kira's largestral: https://rentry.org/kiralargestralprompt — Mistral Large
- CommandRP: https://rentry.org/4y1je_commandrp — Command R/R+
- printerJB: https://rentry.org/printerjb — Claude, GPT
- Q1F V1: https://rentry.org/88fr3yr5 — Deepseek
- Minsk: https://rentry.org/minskhub — Gemini
- AIBrain: https://rentry.org/AiBrainPresets — Gemini
- theatreJB/hometheatreJB: https://rentry.org/rekeddeb#theatrejb — Claude, DeepSeek, Nemotron 70B
- Writing Styles: https://rentry.org/deepstyles — Deepseek
- SillyCards: https://sillycards.co/presets.html — Claude, Deepseek, Gemini, GPT, Nous Hermes, Qwen-Max
Your model's provider/proxy isn't available via Chat Completion in your frontend?
Go back to theIf You Want to Use an Online AI
section to learn how to add it.
Your model wastes time explaining itself before playing its turn?
It means that you are using a reasoning model. This new type of model will always "think" before writing its responses.
This reasoning step shouldn't be visible to you unless you open theThinking...
window above the model's turn.
If it is geting mixed with your bot's actual responses, make sure your frontend is updated to a version that actually supports reasoning models, and that support for them isn't disabled.
In SillyTavern, to find this option, click on theAI Response Formatting
button, the third one with anA
in the top bar, and expand theReasoning
section to enable theAuto-Parse
option.
You will see these pages talking about
Latte
from time to time, it is just a nickname forGPT Latest
.
SillyTavern Resources
Extensions
- EmojiPicker: https://github.com/SillyTavern/Extension-EmojiPicker
- Chat Top Info Bar: https://github.com/SillyTavern/Extension-TopInfoBar
- Input History: https://github.com/LenAnderson/SillyTavern-InputHistory
- Quick Persona: https://github.com/SillyTavern/Extension-QuickPersona
- More Flexible Continues: https://github.com/LenAnderson/SillyTavern-MoreFlexibleContinues
- Rewrite: https://github.com/splitclover/rewrite-extension
- Dialogue Colorizer: https://github.com/XanadusWorks/SillyTavern-Dialogue-Colorizer
- Greetings Placeholder: https://github.com/splitclover/greeting-placeholders
- Timelines: https://github.com/SillyTavern/SillyTavern-Timelines
- Tracker: https://github.com/kaldigo/SillyTavern-Tracker
- Stepped Thinking: https://github.com/cierru/st-stepped-thinking
- LALib: https://github.com/LenAnderson/SillyTavern-LALib
Themes
- Moonlit Echoes: https://github.com/RivelleDays/SillyTavern-MoonlitEchoesTheme
- ST-NoShadowDribbblish: https://github.com/IceFog72/ST-NoShadowDribbblish
Quick Replies
- CharacterProvider's Quick Replies: https://rentry.org/CharacterProvider-Quick-Replies
- Guided Generations: https://github.com/Samueras/Guided-Generations
Novel Roleplaying Setups
- Proper Adventure Gaming With LLMs: https://rentry.co/LLMAdventurersGuide — AI Dungeon-like text-adventure setup. Interesting way to roleplay that is less focused on individual characters.
- SX-3: Character Cards Environment: https://huggingface.co/sphiratrioth666/SX-3_Characters_Environment_SillyTavern — A complex modular system to generate starting messages, swap scenarios, clothes, weather and additional roleplay conditions, using only vanilla SillyTavern.
Learning How To Roleplay
Build Your Basic Knowledge
- Local LLM Glossary: https://gist.github.com/kalomaze/4d74e81c3d19ce45f73fa92df8c9b979 — First we have to make sure that we are all speaking the same language, right?
- LLM Samplers Explained: https://gist.github.com/kalomaze/4473f3f975ff5e5fade06e632498f73e — Quick and digestible read to introduce you to the basic samplers.
- Samplers Settings and You - A Comprehensive Beginner Guide: https://rentry.co/samplersettings — A practical follow-up guide that introduces you to the modern samplers and helps you configure a streamlined sampling setup.
- Your settings are (probably) hurting your model - Why sampler settings matter: https://www.reddit.com/r/LocalLLaMA/comments/17vonjo/your_settings_are_probably_hurting_your_model_why/ — They really are! A little more context on why you want to streamline your sampler settings.
- DRY: A modern repetition penalty that reliably prevents looping: https://github.com/oobabooga/text-generation-webui/pull/5677 — Technical explanation of how the DRY sampler works, if you are curious.
- Exclude Top Choices (XTC): A sampler that boosts creativity, breaks writing clichés, and inhibits non-verbatim repetition: https://github.com/oobabooga/text-generation-webui/pull/6335 — Technical explanation of how the XTC sampler works, if you are curious.
- LLM Samplers Visualized: https://artefact2.github.io/llm-sampling/ — Tool that lets you simulate what you've learned above. Play with the samplers and see how they affect the generated tokens.
- Samplers Settings and You - A Comprehensive Beginner Guide: https://rentry.co/samplersettings — A practical follow-up guide that introduces you to the modern samplers and helps you configure a streamlined sampling setup.
- Further Reading: The following are guides that will teach you how to roleplay, how things really work, and give you tips on how to make your sessions better. If you are more interested in learning how to make your own bots, skip to that section and come back when you want to learn more.
- Sukino Guides & Tips for AI Roleplay: https://rentry.org/Sukino-Guides — Shameless self-promotion here. This page isn't really a structured guide, but a collection of tips and best practices related to AI roleplaying that you can read at your own pace. I recommend that you at least read the sections on how to use your turns and what to do when the AI writes something you don't like.
- How 2 Claude: https://rentry.org/how2claude — Interested in taking a peek behind the curtain? In how all this AI roleplaying wizardry really works? How to fix your annoyances? Then read this! It applies to all AI models, despite the name.
- SillyTavern Docs: https://docs.sillytavern.app/ — Not sure how something works? Don't know what an option is for? Read the docs!
Handy Resources for Botmaking
Botmaking is pretty free-form, almost anything you write will work, and everyone does it a little differently, so don't think you need to follow templates or formats to make good bots, plain text is more than fine...
- Character Creation Guide (+JED Template): https://rentry.co/CharacterProvider-GuideToBotmaking — ...That said, in my opinion, the JED+ template is great for beginners, a nice set of training wheels. It gets your character started by simply filling a character sheet, while remaining flexible enough to accommodate any character concept. Some advice in the guide seems a bit odd, especially on how to write an intro and the premise stuff, but the template itself is good, and you'll find different perspectives from other botmakers in the following guides.
- pixi's practical modern botmaking: https://rentry.org/pixiguide — Succinct guide to introduce you to some botmaking good practices, and to what kind of cards you can make.
- Demystifying The Context; Or Common Botmaking Misconceptions: https://rentry.org/Sukino-Guides#demystifying-the-context-or-common-botmaking-misconceptions — Hey look, it's me with a pretentious title. I think this article turned out pretty good. I give you some good practices and warn you about the pitfalls of botmaking.
- BONER'S BOT BUILDING TIPS: https://rentry.org/Bonersbottips — Still relevant as always. While this guide covers the same ground as mine, it is a classic, and its aggressive teaching methods may work better for you.
- How to Create Lorebooks - by NG: https://rentry.co/SillyT_Lorebook — A quick introduction to Lorebooks/World Info. They are a big step up for when you're ready to make your characters deeper and more complex.
- World Info Encyclopedia: https://rentry.co/world-info-encyclopedia — Learn more in-depth about Lorebooks, and how powerful they are.
- Give Your Characters Memory - A Practical Step-by-Step Guide to Data Bank: Persistent Memory via RAG Implementation: https://www.reddit.com/r/SillyTavernAI/comments/1f2eqm1/give_your_characters_memory_a_practical/ — Probably overkill for most people. This is to make your character have long-term memory. I've never experimented with RAG myself, but this guide at least made me understand what it is...sort of.
- Getting to Know the Other Templates: Again, don't think you need to use these formats to make good bots, they have their use cases, but plain text is more than fine these days. However, even if you don't plan to use them, these guides are still worth reading, as the people who write them have valuable insights into how to make your bots better.
- PList + Ali:Chat: This format was really popular before we got models with big contexts. It maximizes token efficiency by combining Python/JSON-style lists for defining character traits with example dialogues to lock in distinct narration and speech patterns. This dual approach is particularly powerful for keeping established characters true to form, expressing subtle personality traits through dialogue, or handling complicated speech patterns. While plain text descriptions can lead to loose interpretations, PList + Ali:Chat provides precise control over character behavior, and prevents your own writing style from bleeding into the character. Just consider whether the added complexity is worth the benefits for your specific use case.
- Trappu's Bot Guide: https://wikia.schneedc.com/bot-creation/trappu
- MinimALIstic (Ali:Chat Lite): https://rentry.co/kingbri-chara-guide
- How to write in PList (Python list) + Ali:Chat: https://rentry.co/plists_alichat_avakson
- StatuoTW's Guide to Making Bots: https://rentry.co/statuobotmakie
- Ali:Chat Style: https://rentry.co/alichat
- BONER'S ALI:CHAT GUIDE FOR FOR MORONS LIKE ME: https://docs.google.com/document/d/1PmU7-MA25P41Q45yU0CpA66Jra51LI-WI1PwSXn2FMs/edit
- W++: Honestly, this format has no redeeming qualities anymore, it is just an inferior PList — use it instead, or simply Markdown, if you want a structured list. But, as obsolete as it is, you will still see it around, from old cards, and people who still like to use it, so you might want to understand what it does.
- W++ For Dummies: https://rentry.co/WPP_For_Dummies
- Pygmalion Tips: https://rentry.org/pygtips
- Other Templates: Botmakers that shared their own templates.
- Shirohibiki's Bot Creation Template: https://rentry.co/shirohibikis-bot-template
- absolutetrash's Bot Guide and Templates: https://rentry.org/absolutetrashs-bot-guide
- PList + Ali:Chat: This format was really popular before we got models with big contexts. It maximizes token efficiency by combining Python/JSON-style lists for defining character traits with example dialogues to lock in distinct narration and speech patterns. This dual approach is particularly powerful for keeping established characters true to form, expressing subtle personality traits through dialogue, or handling complicated speech patterns. While plain text descriptions can lead to loose interpretations, PList + Ali:Chat provides precise control over character behavior, and prevents your own writing style from bleeding into the character. Just consider whether the added complexity is worth the benefits for your specific use case.
- Prompting:
- JINXBREAKS: https://rentry.org/jinxbreaks — Trying to make a crazy character but can't get it to behave the way you want? Maybe this page can help you get an idea of how to prompt it.
- sphiratrioth666's Character Generation Templates: https://huggingface.co/sphiratrioth666/Character_Generation_Templates — Nothing beats a handcrafted bot. But it's handy to be able to have the AI generate characters for you, perhaps to use as a base, or to quickly roleplay with a pre-existing character. These are prompts to be used on any model of your choice.
- Online Editors:
- AI Character Editor: https://desune.moe/aichared/
- Agnastic's Create a Character: https://agnai.chat/editor
- Sharing:
- Tagging & You: A Guide to Tagging Your Bots on Chub AI https://theunofficialguidetochubai.wordpress.com/2025/01/21/tagging-you-a-guide-to-tagging-your-bots-on-chub-ai/
Other Indexes
More people sharing collections of stuff. Just pay attention to when these guides and resources were created and last updated, they may be outdated or contain outdated practices. A lot of these guides come from a time when AI roleplaying was pretty new and we didn't have advanced models with big context windows, everyone was learning and experimenting with what worked best.
- The meta list of various bot making guides: https://rentry.org/meta_botmaking_list
- Chub Discord’s List of Botmaking Resources: https://rentry.co/botmaking
- Bot-Making Resources for JanitorAI.com: https://rentry.co/jaibotmakingresources
- A list of various Jail Breaks for different models: https://rentry.org/jb-listing
- AICG OP template: https://rentry.org/aicgOP