n4ze3m 9f383a81b6 feat: Add generation info to messages
This commit introduces a new feature that displays generation information for each message in the chat.

The generation info is displayed in a popover and includes details about the model used, the prompt, and other relevant information. This helps users understand how their messages were generated and troubleshoot any issues that may arise.

The generation info is retrieved from the LLM response and is stored in the database alongside other message details.

This commit also includes translations for the generation info label in all supported languages.
2024-11-09 15:17:59 +05:30

122 lines
3.8 KiB
JSON

{
"pageAssist": "Page Assist",
"selectAModel": "Select a Model",
"save": "Save",
"saved": "Saved",
"cancel": "Cancel",
"retry": "Retry",
"share": {
"tooltip": {
"share": "Share"
},
"modal": {
"title": "Share link to Chat"
},
"form": {
"defaultValue": {
"name": "Anonymous",
"title": "Untitled chat"
},
"title": {
"label": "Chat title",
"placeholder": "Enter Chat title",
"required": "Chat title is required"
},
"name": {
"label": "Your name",
"placeholder": "Enter your name",
"required": "Your name is required"
},
"btn": {
"save": "Generate Link",
"saving": "Generating Link..."
}
},
"notification": {
"successGenerate": "Link copied to clipboard",
"failGenerate": "Failed to generate link"
}
},
"copyToClipboard": "Copy to clipboard",
"webSearch": "Searching the web",
"regenerate": "Regenerate",
"edit": "Edit",
"delete": "Delete",
"saveAndSubmit": "Save & Submit",
"editMessage": {
"placeholder": "Type a message..."
},
"submit": "Submit",
"noData": "No data",
"noHistory": "No chat history",
"chatWithCurrentPage": "Chat with current page",
"beta": "Beta",
"tts": "Read aloud",
"currentChatModelSettings": "Current Chat Model Settings",
"modelSettings": {
"label": "Model Settings",
"description": "Set the model options globally for all chats",
"form": {
"keepAlive": {
"label": "Keep Alive",
"help": "controls how long the model will stay loaded into memory following the request (default: 5m)",
"placeholder": "Enter Keep Alive duration (e.g. 5m, 10m, 1h)"
},
"temperature": {
"label": "Temperature",
"placeholder": "Enter Temperature value (e.g. 0.7, 1.0)"
},
"numCtx": {
"label": "Number of Contexts",
"placeholder": "Enter Number of Contexts value (default: 2048)"
},
"seed": {
"label": "Seed",
"placeholder": "Enter Seed value (e.g. 1234)",
"help": "Reproducibility of the model output"
},
"topK": {
"label": "Top K",
"placeholder": "Enter Top K value (e.g. 40, 100)"
},
"topP": {
"label": "Top P",
"placeholder": "Enter Top P value (e.g. 0.9, 0.95)"
},
"numGpu": {
"label": "Num GPU",
"placeholder": "Enter number of layers to send to GPU(s)"
},
"systemPrompt": {
"label": "Temporary System Prompt",
"placeholder": "Enter System Prompt",
"help": "This is a quick way to set the system prompt in the current chat, which will override the selected system prompt if it exists."
}
},
"advanced": "More Model Settings"
},
"copilot": {
"summary": "Summarize",
"explain": "Explain",
"rephrase": "Rephrase",
"translate": "Translate",
"custom": "Custom"
},
"citations": "Citations",
"segmented": {
"ollama": "Ollama Models",
"custom": "Custom Models"
},
"downloadCode": "Download Code",
"date": {
"pinned": "Pinned",
"today": "Today",
"yesterday": "Yesterday",
"last7Days": "Last 7 Days",
"older": "Older"
},
"pin": "Pin",
"unpin": "Unpin",
"generationInfo": "Generation Info"
}