Adds a new setting to control the maximum number of tokens generated by the model. This provides more control over the length of responses and can be useful for limiting the amount of text generated in certain situations.
This commit introduces a new feature that displays generation information for each message in the chat.
The generation info is displayed in a popover and includes details about the model used, the prompt, and other relevant information. This helps users understand how their messages were generated and troubleshoot any issues that may arise.
The generation info is retrieved from the LLM response and is stored in the database alongside other message details.
This commit also includes translations for the generation info label in all supported languages.
Adds a "Save" button to the edit message form in Playground, allowing users to save changes without immediately submitting them. This also introduces a new `isSend` flag to the `onEditFormSubmit` prop, enabling developers to control whether a message should be sent immediately or saved for later submission. This enhances flexibility and user control during the message editing process.
Adds pin/unpin functionality to the chat history sidebar, allowing users to keep important conversations readily accessible. This improves user experience and helps organize past interactions.
This feature includes:
- Pin/unpin buttons in the chat history sidebar.
- Updated database schema to include `is_pinned` field for chat history items.
- Localized translations for pin/unpin actions.
- Updated UI to display pinned items at the top of the list.
Adds a new "Download Code" button to the code block component, allowing users to download the code displayed for offline use.
This feature enhances user convenience and provides a more versatile experience for exploring and utilizing code snippets.
Refine the Playground's UI to improve user experience:
- Streamline chat window layout for better message readability
- Introduce a knowledge selection dropdown for easier context setting
- Improve image upload integration for a smoother workflow
- Optimize spacing and styling for a more polished visual appearance
Adds an HTML preview feature to the code block component, allowing users to view the rendered output of their HTML code snippets. This improves the user experience by providing a more interactive and informative way to understand the code.
Shorten the displayed embedding model names in the knowledge settings by removing the "embedding" suffix. This improves readability and makes the table more concise.
Adds support for Ollama embedding, enabling the use of Ollama as an embedding model for RAG.
This allows users to leverage Ollama's advanced embedding capabilities for better document understanding and retrieval.
Adds model type support for chat and embedding models. This allows users to specify which type of model they want to use when adding custom models.
Additionally, this commit introduces a more descriptive interface for adding custom models, enhancing the clarity of the model selection process.
Update the user interface for the OpenAI settings to be more informative and consistent. Improved language around "OpenAI API Settings" to be more accurate and less technical. Also added a helpful tip for using LM Studio as a provider.
Refine settings layout for better visual flow, enlarge settings sidebar for easier viewing, and adjust model refresh logic for more responsive updates.
Refactor: Add more visual weight to the settings panel and optimize model refresh logic.
Refactor: Enhance the visual appeal of settings panel by increasing its width and improving model refresh behavior for better user experience.
Refactor embedding models and their handling to improve performance and simplify the process.
Add a new model selection mechanism, and enhance the UI for model selection, offering clearer and more user-friendly options for embedding models.
Refactor embeddings to use a common model for page assist and RAG, further improving performance and streamlining the workflow.
When using the LM Studio provider, the model selection modal was incorrectly appearing. This commit prevents the modal from appearing when the provider is set to LM Studio. This ensures that the user experience is consistent across all providers.
Make the loading state of OpenAI models more visually appealing and user-friendly by replacing the default `Spin` component with a more contextually relevant loading animation within a centered container. This provides a better visual cue to users while models are being fetched.
Add a provider selection dropdown to the OpenAI settings, enabling users to choose from pre-configured options like "Azure" or "Custom." This streamlines setup and allows for more flexibility in configuring OpenAI API endpoints. The dropdown pre-populates base URLs and names based on the selected provider.
The dropdown also automatically populates base URLs and names based on the selected provider, further simplifying the configuration process.
Adds the title of the current chat history to the share modal. This makes it easier for users to share their conversations with others, as the title provides context for the conversation.
The title is fetched from the database and displayed in the share modal.
Adds a new setting that allows users to set a temporary system prompt for the current chat.
This prompt will override the selected system prompt if it exists.
The new setting is available in the "Current Chat Model Settings" modal.
This feature provides a way to quickly experiment with different system prompts without having to change the default setting.
Adds support for OpenAI models, allowing users to leverage various OpenAI models directly from the application. This includes custom OpenAI models and OpenAI-specific configurations for seamless integration.
This commit introduces a new UI for managing models within the OpenAI integration. This UI allows users to view, add, and delete OpenAI models associated with their OpenAI providers. It includes functionality to fetch and refresh model lists, as well as to search for specific models. These changes enhance the user experience by offering greater control over their OpenAI model interactions.
This commit also includes improvements to the existing OpenAI configuration UI, enabling users to seamlessly manage multiple OpenAI providers and associated models.
Adds a button to the sidebar drawer for clearing chat history.
The button uses a confirmation prompt to ensure user intent and clears the database entries.
This allows users to reset their chat history for a fresh start.
The button uses a tooltip to provide more information on its function.
Adds LaTeX support to the markdown renderer using `rehype-katex` for math equations.
- Replaces block-level LaTeX delimiters `\[ \]` with `$$ $$`.
- Replaces inline LaTeX delimiters `\( \)` with `$ $`.
- Preprocesses the message before rendering to ensure correct delimiters.
This improves the rendering of markdown messages containing mathematical expressions, enhancing the user experience.
Adds a collapsible section to the playground message that displays citations for the response. This is intended to help users better understand the sources used by the model.