You can set a maximum number of tokens for each response to.
Silly tavern response length. (all other sampling methods are disabled) Web wiki security insights new issue [bug] sillytavern takes >15 seconds to send a request to proxy server with long conversations (>6000 context tokens) #228. Presence penalty should be higher.
Web if you want even more exacting control over the humor used in the q & a session, you can use a very common q & a technique. Be sure to choose carefully if you want to become the champion! If supported by the api, you can.
I keep trying to get a response but the ai cant. Most models have context sizes up to. Web the internal equivalent to response length (tokens) should be set to first chunk (tokens) or next chunks (tokens) while the new option is true.
Web sillytavern doesn't generate the responses. So im wondering what settings are good for fast and good responses? Web 1.5 temp for crazy responses.
Web silly question definition: Web here are my settings for tavern, not necessarily the best. Web whats the best setting for silly tavern?
Ago 'memory' when it comes to ai chat bots is a tricky thing. You shouldn't have to go lower than 3.5k though. The larger the parameter value, the longer the generation time takes.