Im using a knowledgebase…
-
Chat keeps ignoring my questions and re-answering the previous question again. Is there a setting that would be causing this? How do I resolve it?
-
Is there a ‘magic’ Token Lengths setting that provides an optimal output (Max Output Lenth, Max Input Length, and Memory Buffer) where we dont run into token errors but still give large short essay responses?
(Provider Error: Rate limit reached for gpt-4o in organization org-ZmGgbPJ9w80B1O08i5zXeMot on tokens per min (TPM): Limit 30000, Used 10008, Requested 25898. Please try again in 11.812s. Visit https://platform.openai.com/account/rate-limits to learn more.)