feat: support for GPU layer
This commit is contained in:
parent
00735cddad
commit
4ef17ff479
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Enter Top P value (e.g. 0.9, 0.95)"
|
"placeholder": "Enter Top P value (e.g. 0.9, 0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "Enter number of layers to send to GPU(s)"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "More Model Settings"
|
"advanced": "More Model Settings"
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Ingresar el valor de Top P (ej: 0.9, 0.95)"
|
"placeholder": "Ingresar el valor de Top P (ej: 0.9, 0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "Ingrese el número de capas para enviar a la(s) GPU(s)"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "Más Configuraciones del Modelo"
|
"advanced": "Más Configuraciones del Modelo"
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "مقدار Top P را وارد کنید (مثلا 0.9, 0.95)"
|
"placeholder": "مقدار Top P را وارد کنید (مثلا 0.9, 0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "تعداد لایههایی که به GPU(ها) ارسال میشود را وارد کنید"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "تنظیمات بیشتر مدل"
|
"advanced": "تنظیمات بیشتر مدل"
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Entrez la valeur Top P (par exemple 0,9, 0,95)"
|
"placeholder": "Entrez la valeur Top P (par exemple 0,9, 0,95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "Entrez le nombre de couches à envoyer au(x) GPU(s)"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "Plus de paramètres du modèle"
|
"advanced": "Plus de paramètres du modèle"
|
||||||
|
@ -81,7 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Inserisci il Valore Top P (e.g. 0.9, 0.95)"
|
"placeholder": "Inserisci il Valore Top P (e.g. 0.9, 0.95)"
|
||||||
}
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "Inserisci il numero di layer da inviare alla/e GPU" }
|
||||||
},
|
},
|
||||||
"advanced": "Altre Impostazioni del Modello"
|
"advanced": "Altre Impostazioni del Modello"
|
||||||
},
|
},
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Top P値を入力してください(例:0.9、0.95)"
|
"placeholder": "Top P値を入力してください(例:0.9、0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "GPU(s)に送信するレイヤー数を入力してください"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "その他のモデル設定"
|
"advanced": "その他のモデル設定"
|
||||||
|
@ -80,6 +80,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "ടോപ് P",
|
"label": "ടോപ് P",
|
||||||
"placeholder": "ടോപ് P മൂല്യം നൽകുക (ഉദാ: 0.9, 0.95)"
|
"placeholder": "ടോപ് P മൂല്യം നൽകുക (ഉദാ: 0.9, 0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "ജിപിയു എണ്ണം",
|
||||||
|
"placeholder": "ജിപിയു(കൾ)ക്ക് അയക്കേണ്ട ലേയറുകളുടെ എണ്ണം നൽകുക"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "കൂടുതൽ മോഡൽ ക്രമീകരണങ്ങൾ"
|
"advanced": "കൂടുതൽ മോഡൽ ക്രമീകരണങ്ങൾ"
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Digite o valor do Top P (ex: 0.9, 0.95)"
|
"placeholder": "Digite o valor do Top P (ex: 0.9, 0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPUs",
|
||||||
|
"placeholder": "Digite o número de camadas para enviar para a(s) GPU(s)"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "Mais Configurações do Modelo"
|
"advanced": "Mais Configurações do Modelo"
|
||||||
|
@ -81,7 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "Введите значение Top P (например, 0.9, 0.95)"
|
"placeholder": "Введите значение Top P (например, 0.9, 0.95)"
|
||||||
}
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "Введите количество слоев для отправки на GPU" }
|
||||||
},
|
},
|
||||||
"advanced": "Больше настроек модели"
|
"advanced": "Больше настроек модели"
|
||||||
},
|
},
|
||||||
|
@ -81,6 +81,10 @@
|
|||||||
"topP": {
|
"topP": {
|
||||||
"label": "Top P",
|
"label": "Top P",
|
||||||
"placeholder": "输入Top P值(例如:0.9、0.95)"
|
"placeholder": "输入Top P值(例如:0.9、0.95)"
|
||||||
|
},
|
||||||
|
"numGpu": {
|
||||||
|
"label": "Num GPU",
|
||||||
|
"placeholder": "输入要发送到 GPU 的层数"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"advanced": "更多模型设置"
|
"advanced": "更多模型设置"
|
||||||
|
@ -24,7 +24,8 @@ export const CurrentChatModelSettings = ({ open, setOpen }: Props) => {
|
|||||||
topP: cUserSettings.topP ?? data.topP,
|
topP: cUserSettings.topP ?? data.topP,
|
||||||
keepAlive: cUserSettings.keepAlive ?? data.keepAlive,
|
keepAlive: cUserSettings.keepAlive ?? data.keepAlive,
|
||||||
numCtx: cUserSettings.numCtx ?? data.numCtx,
|
numCtx: cUserSettings.numCtx ?? data.numCtx,
|
||||||
seed: cUserSettings.seed
|
seed: cUserSettings.seed,
|
||||||
|
numGpu: cUserSettings.numGpu ?? data.numGpu,
|
||||||
})
|
})
|
||||||
return data
|
return data
|
||||||
},
|
},
|
||||||
@ -118,6 +119,16 @@ export const CurrentChatModelSettings = ({ open, setOpen }: Props) => {
|
|||||||
placeholder={t("modelSettings.form.topP.placeholder")}
|
placeholder={t("modelSettings.form.topP.placeholder")}
|
||||||
/>
|
/>
|
||||||
</Form.Item>
|
</Form.Item>
|
||||||
|
|
||||||
|
<Form.Item
|
||||||
|
name="numGpu"
|
||||||
|
label={t("modelSettings.form.numGpu.label")}>
|
||||||
|
<InputNumber
|
||||||
|
style={{ width: "100%" }}
|
||||||
|
size="large"
|
||||||
|
placeholder={t("modelSettings.form.numGpu.placeholder")}
|
||||||
|
/>
|
||||||
|
</Form.Item>
|
||||||
</React.Fragment>
|
</React.Fragment>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
@ -6,7 +6,6 @@ import { Form, Skeleton, Input, InputNumber, Collapse } from "antd"
|
|||||||
import React from "react"
|
import React from "react"
|
||||||
import { useTranslation } from "react-i18next"
|
import { useTranslation } from "react-i18next"
|
||||||
|
|
||||||
|
|
||||||
export const ModelSettings = () => {
|
export const ModelSettings = () => {
|
||||||
const { t } = useTranslation("common")
|
const { t } = useTranslation("common")
|
||||||
const [form] = Form.useForm()
|
const [form] = Form.useForm()
|
||||||
@ -41,6 +40,7 @@ export const ModelSettings = () => {
|
|||||||
temperature: number
|
temperature: number
|
||||||
topK: number
|
topK: number
|
||||||
topP: number
|
topP: number
|
||||||
|
numGpu: number
|
||||||
}) => {
|
}) => {
|
||||||
Object.entries(values).forEach(([key, value]) => {
|
Object.entries(values).forEach(([key, value]) => {
|
||||||
setModelSetting(key, value)
|
setModelSetting(key, value)
|
||||||
@ -106,6 +106,17 @@ export const ModelSettings = () => {
|
|||||||
placeholder={t("modelSettings.form.topP.placeholder")}
|
placeholder={t("modelSettings.form.topP.placeholder")}
|
||||||
/>
|
/>
|
||||||
</Form.Item>
|
</Form.Item>
|
||||||
|
<Form.Item
|
||||||
|
name="numGpu"
|
||||||
|
label={t("modelSettings.form.numGpu.label")}>
|
||||||
|
<InputNumber
|
||||||
|
style={{ width: "100%" }}
|
||||||
|
size="large"
|
||||||
|
placeholder={t(
|
||||||
|
"modelSettings.form.numGpu.placeholder"
|
||||||
|
)}
|
||||||
|
/>
|
||||||
|
</Form.Item>
|
||||||
</React.Fragment>
|
</React.Fragment>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
@ -122,7 +122,9 @@ export const useMessage = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -248,7 +250,9 @@ export const useMessage = () => {
|
|||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ??
|
currentChatModelSettings?.numCtx ??
|
||||||
userDefaultModelSettings?.numCtx,
|
userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
const response = await questionOllama.invoke(promptForQuestion)
|
const response = await questionOllama.invoke(promptForQuestion)
|
||||||
query = response.content.toString()
|
query = response.content.toString()
|
||||||
@ -442,7 +446,9 @@ export const useMessage = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -645,7 +651,9 @@ export const useMessage = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -718,7 +726,9 @@ export const useMessage = () => {
|
|||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ??
|
currentChatModelSettings?.numCtx ??
|
||||||
userDefaultModelSettings?.numCtx,
|
userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
const response = await questionOllama.invoke(promptForQuestion)
|
const response = await questionOllama.invoke(promptForQuestion)
|
||||||
query = response.content.toString()
|
query = response.content.toString()
|
||||||
@ -890,7 +900,9 @@ export const useMessage = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -932,8 +944,6 @@ export const useMessage = () => {
|
|||||||
let contentToSave = ""
|
let contentToSave = ""
|
||||||
|
|
||||||
try {
|
try {
|
||||||
|
|
||||||
|
|
||||||
const prompt = await getPrompt(messageType)
|
const prompt = await getPrompt(messageType)
|
||||||
let humanMessage = new HumanMessage({
|
let humanMessage = new HumanMessage({
|
||||||
content: [
|
content: [
|
||||||
|
@ -117,7 +117,9 @@ export const useMessageOption = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -190,7 +192,9 @@ export const useMessageOption = () => {
|
|||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ??
|
currentChatModelSettings?.numCtx ??
|
||||||
userDefaultModelSettings?.numCtx,
|
userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
const response = await questionOllama.invoke(promptForQuestion)
|
const response = await questionOllama.invoke(promptForQuestion)
|
||||||
query = response.content.toString()
|
query = response.content.toString()
|
||||||
@ -360,7 +364,9 @@ export const useMessageOption = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -576,7 +582,9 @@ export const useMessageOption = () => {
|
|||||||
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
topP: currentChatModelSettings?.topP ?? userDefaultModelSettings?.topP,
|
||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
currentChatModelSettings?.numCtx ?? userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
|
|
||||||
let newMessage: Message[] = []
|
let newMessage: Message[] = []
|
||||||
@ -665,7 +673,9 @@ export const useMessageOption = () => {
|
|||||||
numCtx:
|
numCtx:
|
||||||
currentChatModelSettings?.numCtx ??
|
currentChatModelSettings?.numCtx ??
|
||||||
userDefaultModelSettings?.numCtx,
|
userDefaultModelSettings?.numCtx,
|
||||||
seed: currentChatModelSettings?.seed
|
seed: currentChatModelSettings?.seed,
|
||||||
|
numGpu:
|
||||||
|
currentChatModelSettings?.numGpu ?? userDefaultModelSettings?.numGpu
|
||||||
})
|
})
|
||||||
const response = await questionOllama.invoke(promptForQuestion)
|
const response = await questionOllama.invoke(promptForQuestion)
|
||||||
query = response.content.toString()
|
query = response.content.toString()
|
||||||
|
@ -126,7 +126,7 @@ export class ChatOllama
|
|||||||
this.mirostatTau = fields.mirostatTau;
|
this.mirostatTau = fields.mirostatTau;
|
||||||
this.numBatch = fields.numBatch;
|
this.numBatch = fields.numBatch;
|
||||||
this.numCtx = fields.numCtx;
|
this.numCtx = fields.numCtx;
|
||||||
this.numGpu = fields.numGpu;
|
this.numGpu = fields.numGpu === null ? undefined : fields.numGpu;
|
||||||
this.numGqa = fields.numGqa;
|
this.numGqa = fields.numGqa;
|
||||||
this.numKeep = fields.numKeep;
|
this.numKeep = fields.numKeep;
|
||||||
this.numPredict = fields.numPredict;
|
this.numPredict = fields.numPredict;
|
||||||
|
@ -9,7 +9,8 @@ export const pageAssistModel = async ({
|
|||||||
topK,
|
topK,
|
||||||
topP,
|
topP,
|
||||||
numCtx,
|
numCtx,
|
||||||
seed
|
seed,
|
||||||
|
numGpu
|
||||||
}: {
|
}: {
|
||||||
model: string
|
model: string
|
||||||
baseUrl: string
|
baseUrl: string
|
||||||
@ -19,6 +20,7 @@ export const pageAssistModel = async ({
|
|||||||
topP?: number
|
topP?: number
|
||||||
numCtx?: number
|
numCtx?: number
|
||||||
seed?: number
|
seed?: number
|
||||||
|
numGpu?: number
|
||||||
}) => {
|
}) => {
|
||||||
switch (model) {
|
switch (model) {
|
||||||
case "chrome::gemini-nano::page-assist":
|
case "chrome::gemini-nano::page-assist":
|
||||||
@ -35,7 +37,8 @@ export const pageAssistModel = async ({
|
|||||||
topP,
|
topP,
|
||||||
numCtx,
|
numCtx,
|
||||||
seed,
|
seed,
|
||||||
model
|
model,
|
||||||
|
numGpu
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user