3 Commits

Author SHA1 Message Date
n4ze3m
fd654cafdb feat: Add max tokens setting for model generations
Adds a new setting to control the maximum number of tokens generated by the model. This provides more control over the length of responses and can be useful for limiting the amount of text generated in certain situations.
2024-11-09 16:56:47 +05:30
n4ze3m
9f383a81b6 feat: Add generation info to messages
This commit introduces a new feature that displays generation information for each message in the chat.

The generation info is displayed in a popover and includes details about the model used, the prompt, and other relevant information. This helps users understand how their messages were generated and troubleshoot any issues that may arise.

The generation info is retrieved from the LLM response and is stored in the database alongside other message details.

This commit also includes translations for the generation info label in all supported languages.
2024-11-09 15:17:59 +05:30
hkjang
f61ce71d06 add Koraen localization 2024-11-06 14:32:36 +09:00