This is an overview of running LLMs locally for Sillytavern. Not so much specific model recommendations, but an overview of how Prompt Manager | docs.ST.app Sillytavern Noob Guide: Intro to Local Models
These presets are designed for SillyTavern. Tip. The preset below is for Chat Completion only. For my older Text Completion presets, visit my HuggingFace. [FEATURE_REQUEST] Support post-history prompt elements as Text Completion Presets : r/SillyTavernAI
Even if chat completion and text completion presets are presets for text completion · Issue #3920 · SillyTavern/SillyTavern. Frowningface/Silly_Tavern_Presets_Database · Hugging Face Best text completion preset to use? : r/SillyTavernAI
As far as Sillytavern, what is the preferred meta for 'Text completion presets?' And for 'Advanced Formatting?' I've had success with the Best Sillytavern settings for LLM - KoboldCPP : r/SillyTavernAI sphiratrioth666/SillyTavern-Presets-Sphiratrioth · Hugging Face
For equivalent settings in Text Completion APIs, use Advanced Formatting. Naming Presets. If a preset shares a name with one Meta has some new string formats for LLaMa3 and I don't see them in any of the string formats in Silly Tavern. Any suggestions would be
It got me thinking about presets in SillyTavern and realizing that even though I thought I had my settings finetuned I think I could do Click on the "Sliders" tab in Silly Tavern UI (AI Response Configuration). Load up my Settings Preset from the Text Completion Presets list. Switch between Marinara's LLM Hub
LLaMa 3: "Text Completion Presets" and correct "Context Template Sorry if this has been asked before, I just got back to sillytavern on mobile with termux, not that I understood it much before,
ChatML Presets. Nitral-AI's Captain X Text Completion Preset (Source) SillyTavern Master File → Download (Added 2:35 AM Sunday, December 22 Sukino/SillyTavern-Settings-and-Presets · Hugging Face
Text Generation Models. These days I just really use bigger models via API. With Deepseek and GLM, it has become much more accessible, and it's hard to