8 Commits

Author SHA1 Message Date
n4ze3m
c8c71f69cc feat: Add SearXNG search provider and settings 2024-12-01 00:23:17 +05:30
n4ze3m
e5e04c3674 feat: Add option to resume last chat when opening Web UI 2024-11-30 19:04:45 +05:30
n4ze3m
92013f3bfc feat(settings): add Ollama connection status check setting
- Add new setting to enable/disable Ollama connection status check
- Update translations for the new setting across all supported languages
2024-11-17 12:26:14 +05:30
n4ze3m
2409ebc75d feat: Add Ollama and Llamafile to dynamic model fetching
Expanded the list of providers for which models are fetched dynamically to include Ollama and Llamafile, removing the need for manual model addition in the user interface for these providers. This simplifies the user experience and ensures users always have access to the latest models without manual intervention.
2024-11-10 15:38:03 +05:30
n4ze3m
f8791a0707 feat: Introduce temporary chat mode
Adds a new "Temporary Chat" mode for quick, non-persistent conversations. The new mode is available in the header bar and will trigger a visually distinct chat experience with a temporary background color. Temporary chats do not save to the chat history and are meant for short, one-off interactions. This feature enhances flexibility and provides a more convenient option for users who need to quickly interact with the AI without committing the conversation to their history.
2024-11-09 19:10:34 +05:30
n4ze3m
fd654cafdb feat: Add max tokens setting for model generations
Adds a new setting to control the maximum number of tokens generated by the model. This provides more control over the length of responses and can be useful for limiting the amount of text generated in certain situations.
2024-11-09 16:56:47 +05:30
n4ze3m
9f383a81b6 feat: Add generation info to messages
This commit introduces a new feature that displays generation information for each message in the chat.

The generation info is displayed in a popover and includes details about the model used, the prompt, and other relevant information. This helps users understand how their messages were generated and troubleshoot any issues that may arise.

The generation info is retrieved from the LLM response and is stored in the database alongside other message details.

This commit also includes translations for the generation info label in all supported languages.
2024-11-09 15:17:59 +05:30
hkjang
f61ce71d06 add Koraen localization 2024-11-06 14:32:36 +09:00