feat: Add model type support

Adds model type support for chat and embedding models. This allows users to specify which type of model they want to use when adding custom models.

Additionally, this commit introduces a more descriptive interface for adding custom models, enhancing the clarity of the model selection process.
This commit is contained in:
n4ze3m 2024-10-13 18:22:16 +05:30
parent 4e04155471
commit ff4473c35b
9 changed files with 277 additions and 43 deletions

View File

@ -31,7 +31,7 @@
"deleteConfirm": "Are you sure you want to delete this provider?", "deleteConfirm": "Are you sure you want to delete this provider?",
"model": { "model": {
"title": "Model List", "title": "Model List",
"subheading": "Please select the models you want to use with this provider.", "subheading": "Please select the chat models you want to use with this provider.",
"success": "Successfully added new models." "success": "Successfully added new models."
}, },
"tipLMStudio": "Page Assist will automatically fetch the models you loaded on LM Studio. You don't need to add them manually." "tipLMStudio": "Page Assist will automatically fetch the models you loaded on LM Studio. You don't need to add them manually."
@ -41,7 +41,8 @@
"updateSuccess": "Provider updated successfully.", "updateSuccess": "Provider updated successfully.",
"delete": "Delete", "delete": "Delete",
"edit": "Edit", "edit": "Edit",
"refetch": "Refech Model List", "newModel": "Add Models to Provider",
"noNewModel": "For LMStudio, we fetch dynamically. No manual addition needed.",
"searchModel": "Search Model", "searchModel": "Search Model",
"selectAll": "Select All", "selectAll": "Select All",
"save": "Save", "save": "Save",
@ -49,6 +50,7 @@
"manageModels": { "manageModels": {
"columns": { "columns": {
"name": "Model Name", "name": "Model Name",
"model_type": "Model Type",
"model_id": "Model ID", "model_id": "Model ID",
"provider": "Provider Name", "provider": "Provider Name",
"actions": "Action" "actions": "Action"
@ -58,7 +60,31 @@
}, },
"confirm": { "confirm": {
"delete": "Are you sure you want to delete this model?" "delete": "Are you sure you want to delete this model?"
},
"modal": {
"title": "Add Custom Model",
"form": {
"name": {
"label": "Model ID",
"placeholder": "llama3.2",
"required": "Model ID is required."
},
"provider": {
"label": "Provider",
"placeholder": "Select provider",
"required": "Provider is required."
},
"type": {
"label": "Model Type"
}
}
} }
}, },
"noModelFound": "No model found. Make sure you have added correct provider with base URL and API key." "noModelFound": "No model found. Make sure you have added correct provider with base URL and API key.",
"radio": {
"chat": "Chat Model",
"embedding": "Embedding Model",
"chatInfo": "is used for chat completion and conversation generation",
"embeddingInfo": "is used for RAG and other semantic search related tasks."
}
} }

View File

@ -0,0 +1,129 @@
import { createModel } from "@/db/models"
import { getAllOpenAIConfig } from "@/db/openai"
import { useMutation, useQuery, useQueryClient } from "@tanstack/react-query"
import { Input, Modal, Form, Select, Radio } from "antd"
import { Loader2 } from "lucide-react"
import { useTranslation } from "react-i18next"
type Props = {
open: boolean
setOpen: (open: boolean) => void
}
export const AddCustomModelModal: React.FC<Props> = ({ open, setOpen }) => {
const { t } = useTranslation(["openai"])
const [form] = Form.useForm()
const queryClient = useQueryClient()
const { data, isPending } = useQuery({
queryKey: ["fetchProviders"],
queryFn: async () => {
const providers = await getAllOpenAIConfig()
return providers.filter((provider) => provider.provider !== "lmstudio")
}
})
const onFinish = async (values: {
model_id: string
model_type: "chat" | "embedding"
provider_id: string
}) => {
await createModel(
values.model_id,
values.model_id,
values.provider_id,
values.model_type
)
return true
}
const { mutate: createModelMutation, isPending: isSaving } = useMutation({
mutationFn: onFinish,
onSuccess: () => {
queryClient.invalidateQueries({
queryKey: ["fetchCustomModels"]
})
queryClient.invalidateQueries({
queryKey: ["fetchModel"]
})
setOpen(false)
form.resetFields()
}
})
return (
<Modal
footer={null}
open={open}
title={t("manageModels.modal.title")}
onCancel={() => setOpen(false)}>
<Form form={form} onFinish={createModelMutation} layout="vertical">
<Form.Item
name="model_id"
label={t("manageModels.modal.form.name.label")}
rules={[
{
required: true,
message: t("manageModels.modal.form.name.required")
}
]}>
<Input
placeholder={t("manageModels.modal.form.name.placeholder")}
size="large"
/>
</Form.Item>
<Form.Item
name="provider_id"
label={t("manageModels.modal.form.provider.label")}
rules={[
{
required: true,
message: t("manageModels.modal.form.provider.required")
}
]}>
<Select
placeholder={t("manageModels.modal.form.provider.placeholder")}
size="large"
loading={isPending}>
{data?.map((provider: any) => (
<Select.Option key={provider.id} value={provider.id}>
{provider.name}
</Select.Option>
))}
</Select>
</Form.Item>
<Form.Item
name="model_type"
label={t("manageModels.modal.form.type.label")}
initialValue="chat"
rules={[
{
required: true,
message: t("manageModels.modal.form.type.required")
}
]}>
<Radio.Group>
<Radio value="chat">{t("radio.chat")}</Radio>
<Radio value="embedding">{t("radio.embedding")}</Radio>
</Radio.Group>
</Form.Item>
<Form.Item>
<button
type="submit"
disabled={isSaving}
className="inline-flex justify-center w-full text-center mt-4 items-center rounded-md border border-transparent bg-black px-2 py-2 text-sm font-medium leading-4 text-white shadow-sm hover:bg-gray-700 focus:outline-none focus:ring-2 focus:ring-indigo-500 focus:ring-offset-2 dark:bg-white dark:text-gray-800 dark:hover:bg-gray-100 dark:focus:ring-gray-500 dark:focus:ring-offset-gray-100 disabled:opacity-50 ">
{!isSaving ? (
t("common:save")
) : (
<Loader2 className="w-5 h-5 animate-spin" />
)}
</button>
</Form.Item>
</Form>
</Modal>
)
}

View File

@ -1,5 +1,5 @@
import { useForm } from "@mantine/form" import { useForm } from "@mantine/form"
import { useMutation } from "@tanstack/react-query" import { useMutation, useQueryClient } from "@tanstack/react-query"
import { Input, Modal, notification } from "antd" import { Input, Modal, notification } from "antd"
import { Download } from "lucide-react" import { Download } from "lucide-react"
import { useTranslation } from "react-i18next" import { useTranslation } from "react-i18next"
@ -11,6 +11,7 @@ type Props = {
export const AddOllamaModelModal: React.FC<Props> = ({ open, setOpen }) => { export const AddOllamaModelModal: React.FC<Props> = ({ open, setOpen }) => {
const { t } = useTranslation(["settings", "common", "openai"]) const { t } = useTranslation(["settings", "common", "openai"])
const queryClient = useQueryClient()
const form = useForm({ const form = useForm({
initialValues: { initialValues: {

View File

@ -1,7 +1,7 @@
import { getAllCustomModels, deleteModel } from "@/db/models" import { getAllCustomModels, deleteModel } from "@/db/models"
import { useStorage } from "@plasmohq/storage/hook" import { useStorage } from "@plasmohq/storage/hook"
import { useQuery, useQueryClient, useMutation } from "@tanstack/react-query" import { useQuery, useQueryClient, useMutation } from "@tanstack/react-query"
import { Skeleton, Table, Tooltip } from "antd" import { Skeleton, Table, Tag, Tooltip } from "antd"
import { Trash2 } from "lucide-react" import { Trash2 } from "lucide-react"
import { useTranslation } from "react-i18next" import { useTranslation } from "react-i18next"
@ -10,7 +10,6 @@ export const CustomModelsTable = () => {
const { t } = useTranslation(["openai", "common"]) const { t } = useTranslation(["openai", "common"])
const queryClient = useQueryClient() const queryClient = useQueryClient()
const { data, status } = useQuery({ const { data, status } = useQuery({
@ -27,7 +26,6 @@ export const CustomModelsTable = () => {
} }
}) })
return ( return (
<div> <div>
<div> <div>
@ -37,16 +35,20 @@ export const CustomModelsTable = () => {
<div className="overflow-x-auto"> <div className="overflow-x-auto">
<Table <Table
columns={[ columns={[
{
title: t("manageModels.columns.name"),
dataIndex: "name",
key: "name"
},
{ {
title: t("manageModels.columns.model_id"), title: t("manageModels.columns.model_id"),
dataIndex: "model_id", dataIndex: "model_id",
key: "model_id" key: "model_id"
}, },
{
title: t("manageModels.columns.model_type"),
dataIndex: "model_type",
render: (txt) => (
<Tag color={txt === "chat" ? "green" : "blue"}>
{t(`radio.${txt}`)}
</Tag>
)
},
{ {
title: t("manageModels.columns.provider"), title: t("manageModels.columns.provider"),
dataIndex: "provider", dataIndex: "provider",

View File

@ -6,11 +6,13 @@ import { useTranslation } from "react-i18next"
import { OllamaModelsTable } from "./OllamaModelsTable" import { OllamaModelsTable } from "./OllamaModelsTable"
import { CustomModelsTable } from "./CustomModelsTable" import { CustomModelsTable } from "./CustomModelsTable"
import { AddOllamaModelModal } from "./AddOllamaModelModal" import { AddOllamaModelModal } from "./AddOllamaModelModal"
import { AddCustomModelModal } from "./AddCustomModelModal"
dayjs.extend(relativeTime) dayjs.extend(relativeTime)
export const ModelsBody = () => { export const ModelsBody = () => {
const [open, setOpen] = useState(false) const [open, setOpen] = useState(false)
const [openAddModelModal, setOpenAddModelModal] = useState(false)
const [segmented, setSegmented] = useState<string>("ollama") const [segmented, setSegmented] = useState<string>("ollama")
const { t } = useTranslation(["settings", "common", "openai"]) const { t } = useTranslation(["settings", "common", "openai"])
@ -26,6 +28,8 @@ export const ModelsBody = () => {
onClick={() => { onClick={() => {
if (segmented === "ollama") { if (segmented === "ollama") {
setOpen(true) setOpen(true)
} else {
setOpenAddModelModal(true)
} }
}} }}
className="inline-flex items-center rounded-md border border-transparent bg-black px-2 py-2 text-md font-medium leading-4 text-white shadow-sm hover:bg-gray-800 focus:outline-none focus:ring-2 focus:ring-indigo-500 focus:ring-offset-2 dark:bg-white dark:text-gray-800 dark:hover:bg-gray-100 dark:focus:ring-gray-500 dark:focus:ring-offset-gray-100 disabled:opacity-50"> className="inline-flex items-center rounded-md border border-transparent bg-black px-2 py-2 text-md font-medium leading-4 text-white shadow-sm hover:bg-gray-800 focus:outline-none focus:ring-2 focus:ring-indigo-500 focus:ring-offset-2 dark:bg-white dark:text-gray-800 dark:hover:bg-gray-100 dark:focus:ring-gray-500 dark:focus:ring-offset-gray-100 disabled:opacity-50">
@ -56,6 +60,11 @@ export const ModelsBody = () => {
</div> </div>
<AddOllamaModelModal open={open} setOpen={setOpen} /> <AddOllamaModelModal open={open} setOpen={setOpen} />
<AddCustomModelModal
open={openAddModelModal}
setOpen={setOpenAddModelModal}
/>
</div> </div>
) )
} }

View File

@ -1,10 +1,12 @@
import { getOpenAIConfigById } from "@/db/openai" import { getOpenAIConfigById } from "@/db/openai"
import { getAllOpenAIModels } from "@/libs/openai" import { getAllOpenAIModels } from "@/libs/openai"
import { useMutation, useQuery } from "@tanstack/react-query" import { useMutation, useQuery, useQueryClient } from "@tanstack/react-query"
import { useTranslation } from "react-i18next" import { useTranslation } from "react-i18next"
import { Checkbox, Input, Spin, message } from "antd" import { Checkbox, Input, Spin, message, Radio } from "antd"
import { useState, useMemo } from "react" import { useState, useMemo } from "react"
import { createManyModels } from "@/db/models" import { createManyModels } from "@/db/models"
import { Popover } from "antd"
import { InfoIcon } from "lucide-react"
type Props = { type Props = {
openaiId: string openaiId: string
@ -15,6 +17,8 @@ export const OpenAIFetchModel = ({ openaiId, setOpenModelModal }: Props) => {
const { t } = useTranslation(["openai"]) const { t } = useTranslation(["openai"])
const [selectedModels, setSelectedModels] = useState<string[]>([]) const [selectedModels, setSelectedModels] = useState<string[]>([])
const [searchTerm, setSearchTerm] = useState("") const [searchTerm, setSearchTerm] = useState("")
const [modelType, setModelType] = useState("chat")
const queryClient = useQueryClient()
const { data, status } = useQuery({ const { data, status } = useQuery({
queryKey: ["openAIConfigs", openaiId], queryKey: ["openAIConfigs", openaiId],
@ -56,7 +60,8 @@ export const OpenAIFetchModel = ({ openaiId, setOpenModelModal }: Props) => {
const payload = models.map((id) => ({ const payload = models.map((id) => ({
model_id: id, model_id: id,
name: filteredModels.find((model) => model.id === id)?.name ?? id, name: filteredModels.find((model) => model.id === id)?.name ?? id,
provider_id: openaiId provider_id: openaiId,
model_type: modelType
})) }))
await createManyModels(payload) await createManyModels(payload)
@ -68,6 +73,9 @@ export const OpenAIFetchModel = ({ openaiId, setOpenModelModal }: Props) => {
mutationFn: onSave, mutationFn: onSave,
onSuccess: () => { onSuccess: () => {
setOpenModelModal(false) setOpenModelModal(false)
queryClient.invalidateQueries({
queryKey: ["fetchModel"]
})
message.success(t("modal.model.success")) message.success(t("modal.model.success"))
} }
}) })
@ -97,6 +105,7 @@ export const OpenAIFetchModel = ({ openaiId, setOpenModelModal }: Props) => {
<p className="text-sm text-gray-500 dark:text-gray-400"> <p className="text-sm text-gray-500 dark:text-gray-400">
{t("modal.model.subheading")} {t("modal.model.subheading")}
</p> </p>
<Input <Input
placeholder={t("searchModel")} placeholder={t("searchModel")}
value={searchTerm} value={searchTerm}
@ -134,6 +143,35 @@ export const OpenAIFetchModel = ({ openaiId, setOpenModelModal }: Props) => {
))} ))}
</div> </div>
</div> </div>
<div className="flex items-center">
<Radio.Group
onChange={(e) => setModelType(e.target.value)}
value={modelType}>
<Radio value="chat">{t("radio.chat")}</Radio>
<Radio value="embedding">{t("radio.embedding")}</Radio>
</Radio.Group>
<Popover
content={
<div>
<p>
<b className="text-gray-800 dark:text-gray-100">
{t("radio.chat")}
</b>{" "}
{t("radio.chatInfo")}
</p>
<p>
<b className="text-gray-800 dark:text-gray-100">
{t("radio.embedding")}
</b>{" "}
{t("radio.embeddingInfo")}
</p>
</div>
}>
<InfoIcon className="ml-2 h-4 w-4 text-gray-500 cursor-pointer" />
</Popover>
</div>
<button <button
onClick={handleSave} onClick={handleSave}
disabled={isSaving} disabled={isSaving}

View File

@ -14,7 +14,13 @@ import {
updateOpenAIConfig updateOpenAIConfig
} from "@/db/openai" } from "@/db/openai"
import { useMutation, useQuery, useQueryClient } from "@tanstack/react-query" import { useMutation, useQuery, useQueryClient } from "@tanstack/react-query"
import { Pencil, Trash2, RotateCwIcon, DownloadIcon, AlertTriangle } from "lucide-react" import {
Pencil,
Trash2,
RotateCwIcon,
DownloadIcon,
AlertTriangle
} from "lucide-react"
import { OpenAIFetchModel } from "./openai-fetch-model" import { OpenAIFetchModel } from "./openai-fetch-model"
import { OAI_API_PROVIDERS } from "@/utils/oai-api-providers" import { OAI_API_PROVIDERS } from "@/utils/oai-api-providers"
@ -149,17 +155,23 @@ export const OpenAIApp = () => {
</button> </button>
</Tooltip> </Tooltip>
<Tooltip title={t("refetch")}> <Tooltip
title={
record.provider !== "lmstudio"
? t("newModel")
: t("noNewModel")
}>
<button <button
className="text-gray-700 dark:text-gray-400" className="text-gray-700 dark:text-gray-400 disabled:opacity-50"
onClick={() => { onClick={() => {
setOpenModelModal(true) setOpenModelModal(true)
setOpenaiId(record.id) setOpenaiId(record.id)
}} }}
disabled={!record.id}> disabled={!record.id || record.provider === "lmstudio"}>
<DownloadIcon className="size-4" /> <DownloadIcon className="size-4" />
</button> </button>
</Tooltip> </Tooltip>
<Tooltip title={t("delete")}> <Tooltip title={t("delete")}>
<button <button
className="text-red-500 dark:text-red-400" className="text-red-500 dark:text-red-400"
@ -251,11 +263,11 @@ export const OpenAIApp = () => {
placeholder={t("modal.apiKey.placeholder")} placeholder={t("modal.apiKey.placeholder")}
/> />
</Form.Item> </Form.Item>
{ {provider === "lmstudio" && (
provider === "lmstudio" && <div className="text-xs text-gray-600 dark:text-gray-400 mb-4"> <div className="text-xs text-gray-600 dark:text-gray-400 mb-4">
{t("modal.tipLMStudio")} {t("modal.tipLMStudio")}
</div> </div>
} )}
<button <button
type="submit" type="submit"
className="inline-flex justify-center w-full text-center mt-4 items-center rounded-md border border-transparent bg-black px-2 py-2 text-sm font-medium leading-4 text-white shadow-sm hover:bg-gray-700 focus:outline-none focus:ring-2 focus:ring-indigo-500 focus:ring-offset-2 dark:bg-white dark:text-gray-800 dark:hover:bg-gray-100 dark:focus:ring-gray-500 dark:focus:ring-offset-gray-100 disabled:opacity-50"> className="inline-flex justify-center w-full text-center mt-4 items-center rounded-md border border-transparent bg-black px-2 py-2 text-sm font-medium leading-4 text-white shadow-sm hover:bg-gray-700 focus:outline-none focus:ring-2 focus:ring-indigo-500 focus:ring-offset-2 dark:bg-white dark:text-gray-800 dark:hover:bg-gray-100 dark:focus:ring-gray-500 dark:focus:ring-offset-gray-100 disabled:opacity-50">

View File

@ -21,18 +21,21 @@ export const generateID = () => {
} }
export const removeModelSuffix = (id: string) => { export const removeModelSuffix = (id: string) => {
return id.replace( return id
/_model-[a-f0-9]{4}-[a-f0-9]{4}-[a-f0-9]{3,4}-[a-f0-9]{4}/, .replace(/_model-[a-f0-9]{4}-[a-f0-9]{4}-[a-f0-9]{3,4}-[a-f0-9]{4}/, "")
"" .replace(/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/, "")
).replace(/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/, "")
} }
export const isLMStudioModel = (model: string) => { export const isLMStudioModel = (model: string) => {
const lmstudioModelRegex = /_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/ const lmstudioModelRegex =
/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/
return lmstudioModelRegex.test(model) return lmstudioModelRegex.test(model)
} }
export const getLMStudioModelId = (model: string): { model_id: string, provider_id: string } => { export const getLMStudioModelId = (
const lmstudioModelRegex = /_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/ model: string
): { model_id: string; provider_id: string } => {
const lmstudioModelRegex =
/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/
const match = model.match(lmstudioModelRegex) const match = model.match(lmstudioModelRegex)
if (match) { if (match) {
const modelId = match[0] const modelId = match[0]
@ -131,7 +134,7 @@ export class ModelDb {
} }
export const createManyModels = async ( export const createManyModels = async (
data: { model_id: string; name: string; provider_id: string }[] data: { model_id: string; name: string; provider_id: string, model_type: string }[]
) => { ) => {
const db = new ModelDb() const db = new ModelDb()
@ -142,7 +145,6 @@ export const createManyModels = async (
id: `${item.model_id}_${generateID()}`, id: `${item.model_id}_${generateID()}`,
db_type: "openai_model", db_type: "openai_model",
name: item.name.replaceAll(/accounts\/[^\/]+\/models\//g, ""), name: item.name.replaceAll(/accounts\/[^\/]+\/models\//g, ""),
model_type: "chat"
} }
}) })
@ -160,7 +162,8 @@ export const createManyModels = async (
export const createModel = async ( export const createModel = async (
model_id: string, model_id: string,
name: string, name: string,
provider_id: string provider_id: string,
model_type: string
) => { ) => {
const db = new ModelDb() const db = new ModelDb()
const id = generateID() const id = generateID()
@ -171,7 +174,7 @@ export const createModel = async (
provider_id, provider_id,
lookup: `${model_id}_${provider_id}`, lookup: `${model_id}_${provider_id}`,
db_type: "openai_model", db_type: "openai_model",
model_type: "chat" model_type: model_type
} }
await db.create(model) await db.create(model)
return model return model
@ -186,9 +189,15 @@ export const getModelInfo = async (id: string) => {
throw new Error("Invalid LMStudio model ID") throw new Error("Invalid LMStudio model ID")
} }
return { return {
model_id: id.replace(/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/, ""), model_id: id.replace(
/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/,
""
),
provider_id: `openai-${lmstudioId.provider_id}`, provider_id: `openai-${lmstudioId.provider_id}`,
name: id.replace(/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/, "") name: id.replace(
/_lmstudio_openai-[a-f0-9]{4}-[a-f0-9]{3}-[a-f0-9]{4}/,
""
)
} }
} }
@ -207,6 +216,7 @@ export const getAllCustomModels = async () => {
return { ...model, provider } return { ...model, provider }
}) })
) )
return modelsWithProvider return modelsWithProvider
} }
@ -247,14 +257,16 @@ export const dynamicFetchLMStudio = async ({
id: `${e?.id}_lmstudio_${providerId}`, id: `${e?.id}_lmstudio_${providerId}`,
provider: providerId, provider: providerId,
lookup: `${e?.id}_${providerId}`, lookup: `${e?.id}_${providerId}`,
provider_id: providerId, provider_id: providerId
} }
}) })
return lmstudioModels return lmstudioModels
} }
export const ollamaFormatAllCustomModels = async () => { export const ollamaFormatAllCustomModels = async (
modelType: "all" | "chat" | "embedding" = "all"
) => {
const [allModles, allProviders] = await Promise.all([ const [allModles, allProviders] = await Promise.all([
getAllCustomModels(), getAllCustomModels(),
getAllOpenAIConfig() getAllOpenAIConfig()
@ -276,7 +288,12 @@ export const ollamaFormatAllCustomModels = async () => {
const lmModels = lmModelsFetch.flat() const lmModels = lmModelsFetch.flat()
// merge allModels and lmModels // merge allModels and lmModels
const allModlesWithLMStudio = [...allModles, ...lmModels] const allModlesWithLMStudio = [
...(modelType !== "all"
? allModles.filter((model) => model.model_type === modelType)
: allModles),
...lmModels
]
const ollamaModels = allModlesWithLMStudio.map((model) => { const ollamaModels = allModlesWithLMStudio.map((model) => {
return { return {

View File

@ -138,7 +138,7 @@ export const getEmbeddingModels = async ({ returnEmpty }: {
}) => { }) => {
try { try {
const ollamaModels = await getAllModels({ returnEmpty }) const ollamaModels = await getAllModels({ returnEmpty })
const customModels = await ollamaFormatAllCustomModels() const customModels = await ollamaFormatAllCustomModels("embedding")
return [ return [
...ollamaModels.map((model) => { ...ollamaModels.map((model) => {
@ -217,7 +217,7 @@ export const fetchChatModels = async ({
}) })
const chromeModel = await getChromeAIModel() const chromeModel = await getChromeAIModel()
const customModels = await ollamaFormatAllCustomModels() const customModels = await ollamaFormatAllCustomModels("chat")
return [ return [
...chatModels, ...chatModels,
@ -234,7 +234,7 @@ export const fetchChatModels = async ({
} }
}) })
const chromeModel = await getChromeAIModel() const chromeModel = await getChromeAIModel()
const customModels = await ollamaFormatAllCustomModels() const customModels = await ollamaFormatAllCustomModels("chat")
return [ return [
...models, ...models,
...chromeModel, ...chromeModel,