Sillytavern memory reddit. A place to discuss the SillyTavern fork of TavernAI.

Sillytavern memory reddit When I say "bot" however, I am referring to the characters that a user can create and tell the AI model to roleplay as, I want mainly something that can have a better memory, more mature and don’t hallucinate so much, I don’t mind waiting a bit between a response or other, since I’m playing while I’m doing others things, but don’t plan to wait for more than 1 minute for a reply 🤣 I run everything locally, for free, on a weak laptop that has a NVIDIA GeForce RTX 2060 (6G of GPU memory). The backend itself is whatever you want to use - either a local AI setup like ooba's or koboldcpp or anything that has an OpenAI-compatible API, or an external API like OpenAI, Claude, Mancer, and some others I haven't even heard of. Nvidia has 2 memory associated with graphics, first the vram, then your shared system memory as video mem. That’s the easiest way to extend your bot’s “memory” of a longer chat. The context is being built and sent from a blank state during every generation, everything that exists outside of the context range is not considered for a generation. AI because chai had terrible memory. Experimentation and fine-tuning may be required to optimize the memory capabilities of the A higher token count of the model itself is better because then the model will remember more, but it will take more of the vram and ram. Plus a very simple workflow . Sillytavern provides more advanced features for things like roleplaying. Or check it out in the SillyTavern is a fork of TavernAI 1. This is me using it for around 5 days added with custom tweaks from all of experienced users and comparing it to C. The extra memory is really just worth it. g. What do you mean by "wipes memory"? Language models don't have any persistent memory, we give them a context that represents character information and chat history to receive the reply based on that. etc. You can make the AI more descriptive about specific things by adding a Describe: the car kind of thing in Author's Note and then starting the AI off His car was . But good job for the tutorial, if it works well for you, that's nice! **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. In the video the guy assumes that I know what this URL or IP adress is, which seems to be already filled into the information when he opens up tavern. Token is like a memory, most of the model have 2048 token limit, which isn't a lot, but thanks to the bloke and superhot your can downlaod models that can support upto 8K tokens, if your pc can Handle this of course, I personly limit it now to 4096 **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. According to the generator settings, you can set this to have a max memory upwards of 32000 tokens! Now, another thing that is amazing is the web interface. The official home of #Supernote lineup on Reddit. com's Reddit Forex Trading Community! Here you can converse about trading ideas, strategies, I used the Midnight Miqu 120b self merge and it was pretty good, using more descriptive language than some of the models I tried off of huggingface but it was noticably less 'intelligent' in terms of spatial memory and reasoning as well as repetition (from what I recall, I didn't test it for that long). Or check it out in the app stores &nbsp; &nbsp; Researchers develop 128Mb STT-MRAM with world's fastest write speed for embedded memory tohoku This level of quantization is perfect for mixtral models, and can fit entirely in 3090 or 4090 memory with 32k context if 4-bit cache is enabled. So what MemGPT is designed to intelligently manage different memory tiers in Large Language Models (LLMs) to effectively provide extended context within the limited context window of the model. I know having a vector database is part of it for long term memory and having an unfiltered LLM is important but I am completely new to all of this. Feel free to point out any mistakes of mine so I can fix it. Why does Vector memory seem to almost never work? Seems to me they recall the correct memory maybe 1 in 4 or even 1 in 6 times. If you add them there (examples on your screenshot), they will be removed when your context is full and therefore freeing up memory for other things. Also has some memory slot for each charakter what will be filled automaticaly. Especially the recently expanded Web Search extra is great, giving the AI access to actual web page content so you get detailed and up-to-date information. 5 for Poe which is what i was referring to, is an AI model. There are settings for temperature, repetition penalty, token allotment, etc. They are less expensive but they have much smaller memory. I am exploring ideas, and experimenting. /r/StableDiffusion is back A place to discuss the SillyTavern fork of TavernAI. Internet Culture (Viral) Amazing A place to discuss the SillyTavern fork of TavernAI. Something Along context aware memory management. For what it's worth, in my experience setting up SillyTavern is more complicated than setting up a local LLM. How can I fix that ? Skip to main content. sillytavern. When loading this model, you’ll need to test it with a conversation + that is 4k tokens to make sure you won’t hit any out of memory errors. I use SillyTavern, although I have no knowledge of what each parameter is, I use Mytholite but 2560 Toks sometimes falls short, memory too, that's why I considered using Venus with 4k, I paid the 5 dollars for Mancer but I ran out of them very quickly, that's **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. The following are memories of previous events that may be relevant: <memories> {{text}} </memories> This hopefully tells the model that the stuff within the XML tags are to be treated as memories. Make sure you update SillyTavern and try again. Oh my! Thank you infinitely for your extremely accurate and helpful response. 1; "unlimited memory", CSCareerQuestions protests in solidarity with the developers who made third party reddit apps. Can write about 4000 word in few hours, but almost a finish product. How do I do the worldbuilding? I can't find ANYTHING about worldbuilding anywhere. Context Size or Context Window is like the AI's memory, GPT-3 has around 4000 and Claude around 9000. at the basic level, the memory is limited by the model's max context size. Or check it out in the app stores as well as applications used to run them, but I suggest using a combination of KoboldCPP and SillyTavern. It basically makes your chats look nicer, let’s you use, wrote, or download whole characters with images, build profiles for different characters for yourself to chat as, it can also connect to summarization plugins, so a summary of important events will be included in each prompt, effectively extending the character’s memory beyond the base token limit. 2. Get the Reddit app Scan this QR code to download the app now. I downloaded r/KoboldAI that executes the LLM model (OPT-2. SillyTavern is a fork of TavernAI 1. A good starting point is Oobabooga with exllama_hf, and one of the GPTQ quantizations of the very new MythaLion model (gptq-4bit-128g-actorder_True if you want it a bit resource light, or gptq-4bit-32g-actorder_True if you want it more "accurate"). The embedding database in Silly Tavern is quite simplistic and only searches in the 'memory' of the current chat by relevance to the current prompt, independent of memory age. for instance. It seemed to me that there was a flurry around adding longterm memory beyond context for llama. 2 model for my The embedding database in Silly Tavern is quite simplistic and only searches in the 'memory' of the current chat by relevance to the current prompt, independent of memory age. When a character is selected, click the book icon. Normally, web interfaces stink, but this has one really cool feature for memory. **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with I wanna try using SillyTavern some more but I'm kinda lost on what's the best API to use, We are Reddit's primary hub for all things modding, Memory bandwidth and capacity of high-end Nvidia consumer GPUs 2. Long term memory. Addons. For me, that also happened, but also it doesn't generate the answer in the SillyTavern server, instead, I only see them on Poe and Termux, but on SillyTavern it keeps loading forever. I've been paying the lowest tier on NovelAi so I get really frustrated with the short-term memory, so If moemate's 13b models have better memory, then it would be fun to try **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. A place to discuss the SillyTavern fork of TavernAI. It's a little more selective in what is stored and aims to return more relevant 'memories' and generating a response from the character that treats it as a memory of the earlier event. Then i will try to create my charakteres inside sillytavern and try to put my story inside sillytavern. 6 or more VRAM use then you are spilling into shared memory and **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. Sillytavern and Sillytavern-extras are running as service so it auto start. It also doesn't access other chats, unless you convert a chat memory into a lorebook. Example Dialogue will be pushed out of memory once your chat starts maxing out the AI's memory. Basically, it stores the whole convo into a database, and retrieves whatever parts are **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. Start by downloading That is why I need an AI model that doesn't struggle so much with memory. So after running all the automated install scripts from the sillytavern website, I've been following a video about how to connect my Ollama LLM to sillytavern. **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with And Seraphina is SFW and a good example of how a complex prompt is implemented in SillyTavern. Log In / Sign Up; or flesh out an event for better one-handed reading, confuse AI with details and get fewer events in memory. So I’ve decided to spin up the Sao10K/L3-8B-Stheno-v3. But, again, they have small memory It seems its not bad. 30B: Excellent, but slow to generate responses (3-5 minutes for a good one). As for context, it's just that. Your prompt and chat will occupy the context; anything past that will be forgotten, but there are things you can do later to help with this. It can sometimes lean towards being too horny in ERP scenarios, but this can be carefully edited to avoid such directions. Discover the elegance of the Supernote, an e-notebook designed for distraction-free writing, reading, and annotating. I heard about Openrouter releasing "mixtral 8x7b instruct 32k context" which is supposed to be excellent and free, but I don't know where to get it and if it is what I am looking for. 💾 Long-Term Memory: Create characters that will remember your dialogs, /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. If your model fits into your vram, it will infer as fas as it can, but if you have cpu layers, or the vram is not enought and it need the shared memory, then the inference will be a lot slover, i think its swapping data between vram and system ram, what is a lot slover. AI also had poor memory but also had filters so I tried to find better. Keep in mind that the character definition takes up a part of the context (the permanent tokens) and a 'small' character will leave more context free for the actual chat/RP. **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with As far as making ST remember things from further back, SillyTavern has a summarization function built in under the extras options (the icon that looks like a stack of boxes) that it will update to include a summary of your recent chat, and include every few replies. I usually use it as a memory. For complex information there is also the World Information (aka Lorebooks). as far i understand this function it correct. I'm actually able to have a very story driven RP with World Info, Group Chat, Chat Memory and ChromaDB, up to 12k tokens, and the AI is This subreddit has been temporarily closed in protest of Reddit's attempt to kill third-party apps **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. I'm not saying this from a hater's perspective, I just fucking wish I find well made bots like Raiden Shogun or Yor Forger from Character. The second thing I do is in the document itself: I start each paragraph with a little explainer, e. The model itself learns nothing, and you can swap models as much as you want without losing anything. This story takes place in California, America. r/PcBuild. Koboldcpp is a hybrid of features you'd find in oobabooga and Sillytavern. Welcome to FXGears. It can replace one or both. Than the lorebook can be included as context information in other chats. Anyway Character description always stays in memory. I don't know about Linode and Vultr. For hardware, I am running a 1080ti 11gb but haven't found much information on how well that runs and what sized models I should be using. Most models have context sizes up to 2048 tokens. Prompt=basic set of initial instructions. SillyTavern is a Hey everyone, Since my last review on Steelskull/L3-Aethora-15B has generated some interest, I've decided to give a smaller 8B model a chance to shine against it. Hey everyone, Since my last review on Steelskull/L3-Aethora-15B has generated some interest, I've decided to give a smaller 8B model a chance to shine against it. The best place for fast help with SillyTavern issues is joining the My system uses two buffers: one is long memory which is the latest output of BART and a short memory which contains the recent chat messages AND the long buffer combined. I just joined after finally figuring out how to download SillyTavern I want to make my own characters but I am confused on how to do it? Get the Reddit app Scan this QR code to download the app now. then after View community ranking In the Top 10% of largest communities on Reddit Is there a better one than "Summarize" I'm wanting to make the AI remember EVERYTHING in the past 25 messages, by doing it by "Summarize" doesn't add the detail. EDIT: time to test out sillytavern and see if it's as easy to set up as oogabooga was. Yep. At the bottom you find a field to enter example messages. (!!!) of GPU memory. For the best results I usually put it two messages deep, which means it is always being put within the memory of the model and it repeats every two messages. Possible even separate module for memory processing, understanding, and working with the pygmalion model, that talk to each other like conscious. I think i will play with the stuff later in the day. Expand user menu Open settings menu. How the fuck can i install ChromaDB onto my phone so i can have long term memory. I've been using ChatGPT. To improve the models 'memory' in very long roleplays, look at the Summarize or Vector Database extensions. Then again maybe this sort of vector storage feature is a base feature of sillytavern? A place to discuss the SillyTavern fork of TavernAI. 8 which is under more active development, I started now with the stage release and the news are amazing for AI memory and knowledge. Even then they usually hallucinate part of the facts surrounding the memory. As for setting it in America, making that explicit in Memory should work. The other two regard World Info (it's like dictionary files), Scan depth is how far back it scans in the conversation for keywords to define You send the prompt via tavern AI, we take the last thing said separate it from the context send it to to another part of the code that handles memories there it does calculations, similarities, adds the prompt then at the same time, looks for a memory resembling the prompt if one exists adds it into the context but it adds it as [this is a memory: ~! memory ~!] context + prompt. Expand If you check task manager and it’s showing 15. You'll be able to see specifically the moment you "spill" over into system memory. A place to discuss the SillyTavern fork of actually a game changer, holy shit. In one of those moments I had an idea There is an extra at SillyTavern/TavernAI that does that automatically. Then, install sillytavern, connect to it, select the llama3 context and instruct (also make sure you use the instruct mode), select universal light preset for samplers and you're ready to start your (e)rp after getting some characters from chub. Open menu Open navigation Go to Reddit Home. So here I am. [Tuesday afternoon, in the A place to discuss the SillyTavern fork of TavernAI. It's a merge of the beloved MythoMax with the very new Pygmalion-2 13B model, and the result is a model that acts a bit better than **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. Some users recommend the use of "Summary" in SillyTavern. EDIT2: Reading the sillytavern installation instructions and apparently to use extras on sillytavern I'll need to get this microsoft build tools nonsense figured out correctly. I make use of all of that and local AI is finally truly useful. I'm not saying this from a hater's perspective, I just fucking ChromaDB from the Silly Tavern extras solves a lot of the memory issues at this tier. if you wish to have an AI with limitless memory it's just development of AI takes years and also Open AI plans to have a model that has 1M in context. Not sure if it's the same thing exactly, but SillyTavern has a similar feature under extras where it injects and keeps a running tab on key summarized events from chat and also a longer term memory as well. The rest you can tweak to your liking. In short, download koboldcpp, download that model in gguf variant and you can already use it. The memory is exactly the same if not worst in my experience, since it has to remember too many unnecesary, annoying walls of texts instead of simple texts like in Charactar. According to the sillytavern extras git, Get the Reddit app Scan this QR code to download the app now. But those are just the "basics" I am looking at adding other techniques into it and into 1 single memory system. It gives you more context in your chats based on past information. Plus, being sparse MoE models they're wicked fast. After some tests I can say that both models are really good for rp, and NoromaidxOpenGPT4-2 is a lot better than older Noromaid versions imo. There's ways to do it, the easiest one is to use the summarize function on ST and then copy it to an always active lorebook Ok. 8 which is under more active development, and has how it performs and compares. There is no END_OF_DIALOG tag in ST Example Dialogue. SillyTavern is a I am in the final stage of testing my Long Term memory and short term memory plugin Memoir+. 8 which is under more active development, (multiple chatting APIs even featuring GPT4 Turbo and Claude 2. Temp makes the bot more creative, although past 1 it tends to get whacky. You don't see it obviously, but its there. (was told this was better than after or insert at set point) **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. ) and will give you responses at full context in about 10 seconds (it also supports token streaming, so you can read each word in **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. I've been trying for a whole week reading up finding a way to get long term memory with my new install of Silly Tavern. I'm really enjoying the interface of SillyTavern, but im really struggling with the actual AI itself. I've found a plethora of install tutorials but NOTHING on what to do after you get everything set up. ST always keeps the Description in the AI's memory. Long term memory strategies? **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. just don't download Stable Diffusion or Eleven Labs with it. 2 model for my Get the Reddit app Scan this QR code to download the app now. let's say you loaded a model, that has 8k context(how much memory the AI can remember), first what you **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. Supernote is a co-design product with our **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. If you care about memory, 2048 on KoboldAI / AI Horde / Oobabooga, the value is ignored for OAI stuff afaik, as they have a much larger, fixed context size. Again, the l2 version of Airoboros is excellent, although the l1 versions of Airochronos and Chronoboros can also do well. ChatGPT, specifically ChatGPT 3. At this point they can be thought of as completely Memory bandwidth and capacity of high-end Nvidia consumer GPUs **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. ai. Personally I've made an install on a server I rent with Online. It was a steep learning curve but I finally got SillyTavern set up with Kobold so it can be run locally. Limited to 8k memory length (due to it's an L3 model). AI. They promote their "ares 70B" model for $20/month. If you want good memory, you can try turbo gpt or claude. It also doesn't access other chats, unless you convert a chat A place to discuss the SillyTavern fork of TavernAI. Click that and at the bottom center it shows you your dedicated graphics memory and then the extra system graphics memory after it. Sillytavern is a frontend. Or check it out in the app stores Due-Memory-6957 SillyTavern is a fork of TavernAI 1. it doesn't give you more memory. Get app Get the Reddit app Log In Log in to Reddit. ai to use in SillyTavern with Poe. SillyTavern extras api has summary and Install Extras and enable vector storage. It can't run LLMs directly, but it can connect to a backend API such as oobabooga. The RTX 3090 has 24GB VRAM, which is plenty of memory to run the newer 13B llama 2 models or slightly older (and slightly better IMO) 30B llama 1 uncensored models (like wizardLM-30B-uncensored-supercot-storytelling which is my personal favorite. Also, damn, most of my clients only know me as Wolfram Ravenwolf, the guy from Reddit and Twitter/X with a wolf and raven as his avatar. Alternatively, does a configuration exist that preserves and extends the memory of especially long chats with /r/StableDiffusion is back open after the protest of Reddit killing open API access To add things permanently, look at the Authors Note or just write it directly into the advanced formatting textbox. Say we move from one location to another I will simply put "{{user}} and {{char}} are at the beach" or whatever. /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, Get app Get the Reddit app Log In Log in to Reddit. 8 which is under more active development, and has added many major I'm a new user and managed to get KoboldCPP and SillyTavern up and Open menu Open navigation Go to Reddit Home. The LLM can access this external memory during inference to enhance the context awareness. This is Reddit’s little corner for iPhone lovers (and some people who just **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. I have been working on a long term memory module for oobabooga/text-generation-webui, I am finally at the point that I have a stable release and could use more help testing. Or check it out in the app stores Memory bandwidth and capacity of high-end Nvidia consumer GPUs 2. I then learned that C. The larger the context you give to the model, the more overhead you need, so it’s possible to have it work just fine with a given set of settings when you start your conversation, but have it run out of VRAM after the conversation gets bigger. Let me ask you one more thing about vectordb: now, I don't know if it's, uhm, a placebo effect or something, but yesterday I used it for the first time after reaching the maximum limit/yellow line, and by clicking on "vectorize all", I noticed an incredible increase in the bot's memory (or something like that). Or check it out in the app stores &nbsp; &nbsp; TOPICS. :-) Also, remove any custom settings, jailbreaks, and all that stuff you still have in SillyTavern from previous online models. More posts you may like r/PcBuild. I've been using SillyTavern for nearly two months now, and I use it exclusively for a chatbot. A community to discuss about large language models for roleplay and writing and the PygmalionAI project - an open-source conversational language model. If you see it in an imported character, delete it or it might confuse the AI. I just got back into it 'memory' when it comes to AI chat bots is a tricky thing. So I'm asking you, Reddit. Apologies if my wording was confusing. SillyTavern itself is fully local, so you can run it on your own computer, even on your mobile phone, since it needs little resources. The main goal of the system is that it uses an internal Ego persona to record the summaries of the conversation as they are happening, then recalls them in a vector database query during chat. . The rest is automatic. it seems like a more advanced version of the summary feature in sillytavern Reply reply Top 6% Rank by size . People who get offended that easily by something so innocent wouldn't be clients I'd like to work with anyway. For the past hour or so I've been trying to install ChromaDB on my phone and well nothing. ), I don't think it would be worth converting into SillyTavern unless you plan on using a larger LLM, taking the time to setup Stable Diffusion (For images), or want to completely switch to chatbot versus **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. r/Stunfisk is your reddit source for news, analyses, and competitive discussion for Pokémon VGC, The model does not store conversations- SillyTavern, Kobold, or whatever program you use to talk to the model stores the conversations. From all the ways I've read so far, thanks to FieldProgrammable, the We're getting there in a few years. And if you want to use 13b models you can run them with Google Collab, it's free. **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with **So What is SillyTavern?** Tavern is a user interface you can install on your computer (and Android phones) that allows you to interact text generation AIs and chat/roleplay with characters you or the community create. You need thé complexe memory , and simple memory. 8 which is under more active development, I'm kinda new to reddit and English is not my main language. r/SillyTavernAI A chip A close button. The answers I obtain aren't bad, just the first message is a repeatition of the initial context, but after a second message, it actually answers, so I don't mind it. Sillytavern, hands down, a great rp local ai. When a I haven't touched sillytavern in a long time (last time I did it was when Poe is still around and is the most used one). Currently a character's memory is basically context-dependent, which makes them memorize a limited number and length of conversations, and the character's setting is not I always run out of memory after a few messages when I use SillyTavern. I have it set to insert before the prompt. 7B-Nerybus-Mix), some ST Extras, plus the latest version of ST that connects back to KoboldAI. Here are things I'll You can find a lot of information for common issues in the SillyTavern Docs: https://docs. Instead of storing the entire chat history and trying to match something relevant in chat, I am going for a kind of synthetic memory. Context=basically its short term memory; try something like 4-8k tokens to start with. Perhaps there are other things I can do to make sure characters don't forget important facts. How/where do I do memory, author's notes and all that? Don't put or leave "extra description" in the Example Dialogue field. I'm using it via Infermatic's API, and perhaps they will extend its memory length in the future (maybe, I don't know—if they do, this model would have almost no flaws). I use a modified version of the latter that also works as a system for adding sillytavern-esque lorebooks to characters. app/. They’re the ones managing the memory, no need to worry about it. MistralAI's own PR said Mixtral requires 100gb, so the open source madness helped them make their product reach farther, too. I've looked up tutorials online but they are all PC related. Was a greatful tip. While I don't often use the text adventure mode for NovelAI (last time I used it was with Sigurd. 8 which is under more active development, and has added many major features. If so, pop open Task Manager and click the "Performance" tab when you load a model. net. Look at the bottom left and you'll see your GPU. Adjust to taste. Like giving it auto-updating date and time, letting it search the web, persistent memory, and voice chat if you want. But they are expensive You can use smaller models with sites like open router or mancer. reddit's new API changes kill third party apps that offer So, I started off with chai then tried out C. or try A place to discuss the SillyTavern fork of TavernAI. It's important to note that the effectiveness of these approaches may vary depending on the specific characteristics of the LLM model and the available resources. the basicAuth wasn't working for me for an unknown reason so I've put other protections. kzf xcvrry lbma slsvi bph sumupn uqjxt fbdyv gluio eaiuz