diff --git a/Seanium_Brain.py b/Seanium_Brain.py index d1e08ca..e0533e6 100644 --- a/Seanium_Brain.py +++ b/Seanium_Brain.py @@ -59,7 +59,7 @@ with st.sidebar: help=_('What sampling temperature to use, between 0 and 1. Higher values like 0.8 will make the ' 'output more random, while lower values like 0.2 will make it more focused and ' 'deterministic. \n\nIt is generally recommend altering this or `top_p` but not both.')) - max_tokens = st.slider(_('Max Tokens'), 850, 4096, value=util.read_json_at(INFO.BRAIN_MEMO, 'max_tokens', 1000), + max_tokens = st.slider(_('Max Tokens'), 10, 4096, value=util.read_json_at(INFO.BRAIN_MEMO, 'max_tokens', 1000), help=_("The maximum number of tokens to generate in the completion.\n\nThe token count of " "your prompt plus `max_tokens` cannot exceed the model's context length. Most " "models have a context length of 2048 tokens (except for the newest models, "