Spaces:
Running
Running
<script lang="ts"> | |
import { v4 as uuidv4 } from 'uuid'; | |
import { toast } from 'svelte-sonner'; | |
import mermaid from 'mermaid'; | |
import { getContext, onMount, tick } from 'svelte'; | |
import { goto } from '$app/navigation'; | |
import { page } from '$app/stores'; | |
import type { Writable } from 'svelte/store'; | |
import type { i18n as i18nType } from 'i18next'; | |
import { OLLAMA_API_BASE_URL, OPENAI_API_BASE_URL, WEBUI_BASE_URL } from '$lib/constants'; | |
import { | |
chatId, | |
chats, | |
config, | |
type Model, | |
models, | |
settings, | |
showSidebar, | |
tags as _tags, | |
WEBUI_NAME, | |
banners, | |
user, | |
socket, | |
showCallOverlay, | |
tools | |
} from '$lib/stores'; | |
import { | |
convertMessagesToHistory, | |
copyToClipboard, | |
extractSentencesForAudio, | |
getUserPosition, | |
promptTemplate, | |
splitStream | |
} from '$lib/utils'; | |
import { generateChatCompletion } from '$lib/apis/ollama'; | |
import { | |
addTagById, | |
createNewChat, | |
deleteTagById, | |
getAllChatTags, | |
getChatById, | |
getChatList, | |
getTagsById, | |
updateChatById | |
} from '$lib/apis/chats'; | |
import { generateOpenAIChatCompletion } from '$lib/apis/openai'; | |
import { runWebSearch } from '$lib/apis/rag'; | |
import { createOpenAITextStream } from '$lib/apis/streaming'; | |
import { queryMemory } from '$lib/apis/memories'; | |
import { getAndUpdateUserLocation, getUserSettings } from '$lib/apis/users'; | |
import { chatCompleted, generateTitle, generateSearchQuery, chatAction } from '$lib/apis'; | |
import Banner from '../common/Banner.svelte'; | |
import MessageInput from '$lib/components/chat/MessageInput.svelte'; | |
import Messages from '$lib/components/chat/Messages.svelte'; | |
import Navbar from '$lib/components/layout/Navbar.svelte'; | |
import CallOverlay from './MessageInput/CallOverlay.svelte'; | |
import { error } from '@sveltejs/kit'; | |
import ChatControls from './ChatControls.svelte'; | |
import EventConfirmDialog from '../common/ConfirmDialog.svelte'; | |
const i18n: Writable<i18nType> = getContext('i18n'); | |
export let chatIdProp = ''; | |
let loaded = false; | |
const eventTarget = new EventTarget(); | |
let showControls = false; | |
let stopResponseFlag = false; | |
let autoScroll = true; | |
let processing = ''; | |
let messagesContainerElement: HTMLDivElement; | |
let showEventConfirmation = false; | |
let eventConfirmationTitle = ''; | |
let eventConfirmationMessage = ''; | |
let eventConfirmationInput = false; | |
let eventConfirmationInputPlaceholder = ''; | |
let eventCallback = null; | |
let showModelSelector = true; | |
let selectedModels = ['']; | |
let atSelectedModel: Model | undefined; | |
let selectedModelIds = []; | |
$: selectedModelIds = atSelectedModel !== undefined ? [atSelectedModel.id] : selectedModels; | |
let selectedToolIds = []; | |
let webSearchEnabled = false; | |
let chat = null; | |
let tags = []; | |
let title = ''; | |
let prompt = ''; | |
let chatFiles = []; | |
let files = []; | |
let messages = []; | |
let history = { | |
messages: {}, | |
currentId: null | |
}; | |
let params = {}; | |
let valves = {}; | |
$: if (history.currentId !== null) { | |
let _messages = []; | |
let currentMessage = history.messages[history.currentId]; | |
while (currentMessage !== null) { | |
_messages.unshift({ ...currentMessage }); | |
currentMessage = | |
currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null; | |
} | |
messages = _messages; | |
} else { | |
messages = []; | |
} | |
$: if (chatIdProp) { | |
(async () => { | |
console.log(chatIdProp); | |
if (chatIdProp && (await loadChat())) { | |
await tick(); | |
loaded = true; | |
window.setTimeout(() => scrollToBottom(), 0); | |
const chatInput = document.getElementById('chat-textarea'); | |
chatInput?.focus(); | |
} else { | |
await goto('/'); | |
} | |
})(); | |
} | |
const chatEventHandler = async (event, cb) => { | |
if (event.chat_id === $chatId) { | |
await tick(); | |
console.log(event); | |
let message = history.messages[event.message_id]; | |
const type = event?.data?.type ?? null; | |
const data = event?.data?.data ?? null; | |
if (type === 'status') { | |
if (message?.statusHistory) { | |
message.statusHistory.push(data); | |
} else { | |
message.statusHistory = [data]; | |
} | |
} else if (type === 'citation') { | |
if (message?.citations) { | |
message.citations.push(data); | |
} else { | |
message.citations = [data]; | |
} | |
} else if (type === 'message') { | |
message.content += data.content; | |
} else if (type === 'replace') { | |
message.content = data.content; | |
} else if (type === 'confirmation') { | |
eventCallback = cb; | |
eventConfirmationInput = false; | |
showEventConfirmation = true; | |
eventConfirmationTitle = data.title; | |
eventConfirmationMessage = data.message; | |
} else if (type === 'input') { | |
eventCallback = cb; | |
eventConfirmationInput = true; | |
showEventConfirmation = true; | |
eventConfirmationTitle = data.title; | |
eventConfirmationMessage = data.message; | |
eventConfirmationInputPlaceholder = data.placeholder; | |
} else { | |
console.log('Unknown message type', data); | |
} | |
messages = messages; | |
} | |
}; | |
onMount(async () => { | |
const onMessageHandler = async (event) => { | |
if (event.origin === window.origin) { | |
// Replace with your iframe's origin | |
console.log('Message received from iframe:', event.data); | |
if (event.data.type === 'input:prompt') { | |
console.log(event.data.text); | |
const inputElement = document.getElementById('chat-textarea'); | |
if (inputElement) { | |
prompt = event.data.text; | |
inputElement.focus(); | |
} | |
} | |
if (event.data.type === 'action:submit') { | |
console.log(event.data.text); | |
if (prompt !== '') { | |
await tick(); | |
submitPrompt(prompt); | |
} | |
} | |
if (event.data.type === 'input:prompt:submit') { | |
console.log(event.data.text); | |
if (prompt !== '') { | |
await tick(); | |
submitPrompt(event.data.text); | |
} | |
} | |
} | |
}; | |
window.addEventListener('message', onMessageHandler); | |
$socket.on('chat-events', chatEventHandler); | |
if (!$chatId) { | |
chatId.subscribe(async (value) => { | |
if (!value) { | |
await initNewChat(); | |
} | |
}); | |
} else { | |
if (!($settings.saveChatHistory ?? true)) { | |
await goto('/'); | |
} | |
} | |
return () => { | |
window.removeEventListener('message', onMessageHandler); | |
$socket.off('chat-events'); | |
}; | |
}); | |
////////////////////////// | |
// Web functions | |
////////////////////////// | |
const initNewChat = async () => { | |
window.history.replaceState(history.state, '', `/`); | |
await chatId.set(''); | |
autoScroll = true; | |
title = ''; | |
messages = []; | |
history = { | |
messages: {}, | |
currentId: null | |
}; | |
chatFiles = []; | |
params = {}; | |
if ($page.url.searchParams.get('models')) { | |
selectedModels = $page.url.searchParams.get('models')?.split(','); | |
} else if ($settings?.models) { | |
selectedModels = $settings?.models; | |
} else if ($config?.default_models) { | |
console.log($config?.default_models.split(',') ?? ''); | |
selectedModels = $config?.default_models.split(','); | |
} else { | |
selectedModels = ['']; | |
} | |
if ($page.url.searchParams.get('q')) { | |
prompt = $page.url.searchParams.get('q') ?? ''; | |
if (prompt) { | |
await tick(); | |
submitPrompt(prompt); | |
} | |
} | |
selectedModels = selectedModels.map((modelId) => | |
$models.map((m) => m.id).includes(modelId) ? modelId : '' | |
); | |
const userSettings = await getUserSettings(localStorage.token); | |
if (userSettings) { | |
settings.set(userSettings.ui); | |
} else { | |
settings.set(JSON.parse(localStorage.getItem('settings') ?? '{}')); | |
} | |
const chatInput = document.getElementById('chat-textarea'); | |
setTimeout(() => chatInput?.focus(), 0); | |
}; | |
const loadChat = async () => { | |
chatId.set(chatIdProp); | |
chat = await getChatById(localStorage.token, $chatId).catch(async (error) => { | |
await goto('/'); | |
return null; | |
}); | |
if (chat) { | |
tags = await getTags(); | |
const chatContent = chat.chat; | |
if (chatContent) { | |
console.log(chatContent); | |
selectedModels = | |
(chatContent?.models ?? undefined) !== undefined | |
? chatContent.models | |
: [chatContent.models ?? '']; | |
history = | |
(chatContent?.history ?? undefined) !== undefined | |
? chatContent.history | |
: convertMessagesToHistory(chatContent.messages); | |
title = chatContent.title; | |
const userSettings = await getUserSettings(localStorage.token); | |
if (userSettings) { | |
await settings.set(userSettings.ui); | |
} else { | |
await settings.set(JSON.parse(localStorage.getItem('settings') ?? '{}')); | |
} | |
params = chatContent?.params ?? {}; | |
chatFiles = chatContent?.files ?? []; | |
autoScroll = true; | |
await tick(); | |
if (messages.length > 0) { | |
history.messages[messages.at(-1).id].done = true; | |
} | |
await tick(); | |
return true; | |
} else { | |
return null; | |
} | |
} | |
}; | |
const scrollToBottom = async () => { | |
await tick(); | |
if (messagesContainerElement) { | |
messagesContainerElement.scrollTop = messagesContainerElement.scrollHeight; | |
} | |
}; | |
const createMessagesList = (responseMessageId) => { | |
const message = history.messages[responseMessageId]; | |
if (message.parentId) { | |
return [...createMessagesList(message.parentId), message]; | |
} else { | |
return [message]; | |
} | |
}; | |
const chatCompletedHandler = async (chatId, modelId, responseMessageId, messages) => { | |
await mermaid.run({ | |
querySelector: '.mermaid' | |
}); | |
const res = await chatCompleted(localStorage.token, { | |
model: modelId, | |
messages: messages.map((m) => ({ | |
id: m.id, | |
role: m.role, | |
content: m.content, | |
info: m.info ? m.info : undefined, | |
timestamp: m.timestamp | |
})), | |
chat_id: chatId, | |
session_id: $socket?.id, | |
id: responseMessageId | |
}).catch((error) => { | |
toast.error(error); | |
messages.at(-1).error = { content: error }; | |
return null; | |
}); | |
if (res !== null) { | |
// Update chat history with the new messages | |
for (const message of res.messages) { | |
history.messages[message.id] = { | |
...history.messages[message.id], | |
...(history.messages[message.id].content !== message.content | |
? { originalContent: history.messages[message.id].content } | |
: {}), | |
...message | |
}; | |
} | |
} | |
if ($chatId == chatId) { | |
if ($settings.saveChatHistory ?? true) { | |
chat = await updateChatById(localStorage.token, chatId, { | |
models: selectedModels, | |
messages: messages, | |
history: history, | |
params: params, | |
files: chatFiles | |
}); | |
await chats.set(await getChatList(localStorage.token)); | |
} | |
} | |
}; | |
const chatActionHandler = async (chatId, actionId, modelId, responseMessageId) => { | |
const res = await chatAction(localStorage.token, actionId, { | |
model: modelId, | |
messages: messages.map((m) => ({ | |
id: m.id, | |
role: m.role, | |
content: m.content, | |
info: m.info ? m.info : undefined, | |
timestamp: m.timestamp | |
})), | |
chat_id: chatId, | |
session_id: $socket?.id, | |
id: responseMessageId | |
}).catch((error) => { | |
toast.error(error); | |
messages.at(-1).error = { content: error }; | |
return null; | |
}); | |
if (res !== null) { | |
// Update chat history with the new messages | |
for (const message of res.messages) { | |
history.messages[message.id] = { | |
...history.messages[message.id], | |
...(history.messages[message.id].content !== message.content | |
? { originalContent: history.messages[message.id].content } | |
: {}), | |
...message | |
}; | |
} | |
} | |
if ($chatId == chatId) { | |
if ($settings.saveChatHistory ?? true) { | |
chat = await updateChatById(localStorage.token, chatId, { | |
models: selectedModels, | |
messages: messages, | |
history: history, | |
params: params, | |
files: chatFiles | |
}); | |
await chats.set(await getChatList(localStorage.token)); | |
} | |
} | |
}; | |
const getChatEventEmitter = async (modelId: string, chatId: string = '') => { | |
return setInterval(() => { | |
$socket?.emit('usage', { | |
action: 'chat', | |
model: modelId, | |
chat_id: chatId | |
}); | |
}, 1000); | |
}; | |
////////////////////////// | |
// Chat functions | |
////////////////////////// | |
const submitPrompt = async (userPrompt, { _raw = false } = {}) => { | |
let _responses = []; | |
console.log('submitPrompt', $chatId); | |
selectedModels = selectedModels.map((modelId) => | |
$models.map((m) => m.id).includes(modelId) ? modelId : '' | |
); | |
if (selectedModels.includes('')) { | |
toast.error($i18n.t('Model not selected')); | |
} else if (messages.length != 0 && messages.at(-1).done != true) { | |
// Response not done | |
console.log('wait'); | |
} else if (messages.length != 0 && messages.at(-1).error) { | |
// Error in response | |
toast.error( | |
$i18n.t( | |
`Oops! There was an error in the previous response. Please try again or contact admin.` | |
) | |
); | |
} else if ( | |
files.length > 0 && | |
files.filter((file) => file.type !== 'image' && file.status !== 'processed').length > 0 | |
) { | |
// Upload not done | |
toast.error( | |
$i18n.t( | |
`Oops! Hold tight! Your files are still in the processing oven. We're cooking them up to perfection. Please be patient and we'll let you know once they're ready.` | |
) | |
); | |
} else { | |
// Reset chat input textarea | |
const chatTextAreaElement = document.getElementById('chat-textarea'); | |
if (chatTextAreaElement) { | |
chatTextAreaElement.value = ''; | |
chatTextAreaElement.style.height = ''; | |
} | |
const _files = JSON.parse(JSON.stringify(files)); | |
chatFiles.push(..._files.filter((item) => ['doc', 'file', 'collection'].includes(item.type))); | |
chatFiles = chatFiles.filter( | |
// Remove duplicates | |
(item, index, array) => | |
array.findIndex((i) => JSON.stringify(i) === JSON.stringify(item)) === index | |
); | |
files = []; | |
prompt = ''; | |
// Create user message | |
let userMessageId = uuidv4(); | |
let userMessage = { | |
id: userMessageId, | |
parentId: messages.length !== 0 ? messages.at(-1).id : null, | |
childrenIds: [], | |
role: 'user', | |
content: userPrompt, | |
files: _files.length > 0 ? _files : undefined, | |
timestamp: Math.floor(Date.now() / 1000), // Unix epoch | |
models: selectedModels.filter((m, mIdx) => selectedModels.indexOf(m) === mIdx) | |
}; | |
// Add message to history and Set currentId to messageId | |
history.messages[userMessageId] = userMessage; | |
history.currentId = userMessageId; | |
// Append messageId to childrenIds of parent message | |
if (messages.length !== 0) { | |
history.messages[messages.at(-1).id].childrenIds.push(userMessageId); | |
} | |
// Wait until history/message have been updated | |
await tick(); | |
_responses = await sendPrompt(userPrompt, userMessageId, { newChat: true }); | |
} | |
return _responses; | |
}; | |
const sendPrompt = async (prompt, parentId, { modelId = null, newChat = false } = {}) => { | |
let _responses = []; | |
// If modelId is provided, use it, else use selected model | |
let selectedModelIds = modelId | |
? [modelId] | |
: atSelectedModel !== undefined | |
? [atSelectedModel.id] | |
: selectedModels; | |
// Create response messages for each selected model | |
const responseMessageIds = {}; | |
for (const modelId of selectedModelIds) { | |
const model = $models.filter((m) => m.id === modelId).at(0); | |
if (model) { | |
let responseMessageId = uuidv4(); | |
let responseMessage = { | |
parentId: parentId, | |
id: responseMessageId, | |
childrenIds: [], | |
role: 'assistant', | |
content: '', | |
model: model.id, | |
modelName: model.name ?? model.id, | |
userContext: null, | |
timestamp: Math.floor(Date.now() / 1000) // Unix epoch | |
}; | |
// Add message to history and Set currentId to messageId | |
history.messages[responseMessageId] = responseMessage; | |
history.currentId = responseMessageId; | |
// Append messageId to childrenIds of parent message | |
if (parentId !== null) { | |
history.messages[parentId].childrenIds = [ | |
...history.messages[parentId].childrenIds, | |
responseMessageId | |
]; | |
} | |
responseMessageIds[modelId] = responseMessageId; | |
} | |
} | |
await tick(); | |
// Create new chat if only one message in messages | |
if (newChat && messages.length == 2) { | |
if ($settings.saveChatHistory ?? true) { | |
chat = await createNewChat(localStorage.token, { | |
id: $chatId, | |
title: $i18n.t('New Chat'), | |
models: selectedModels, | |
system: $settings.system ?? undefined, | |
params: params, | |
messages: messages, | |
history: history, | |
tags: [], | |
timestamp: Date.now() | |
}); | |
await chats.set(await getChatList(localStorage.token)); | |
await chatId.set(chat.id); | |
} else { | |
await chatId.set('local'); | |
} | |
await tick(); | |
} | |
const _chatId = JSON.parse(JSON.stringify($chatId)); | |
await Promise.all( | |
selectedModelIds.map(async (modelId) => { | |
console.log('modelId', modelId); | |
const model = $models.filter((m) => m.id === modelId).at(0); | |
if (model) { | |
// If there are image files, check if model is vision capable | |
const hasImages = messages.some((message) => | |
message.files?.some((file) => file.type === 'image') | |
); | |
if (hasImages && !(model.info?.meta?.capabilities?.vision ?? true)) { | |
toast.error( | |
$i18n.t('Model {{modelName}} is not vision capable', { | |
modelName: model.name ?? model.id | |
}) | |
); | |
} | |
let responseMessageId = responseMessageIds[modelId]; | |
let responseMessage = history.messages[responseMessageId]; | |
let userContext = null; | |
if ($settings?.memory ?? false) { | |
if (userContext === null) { | |
const res = await queryMemory(localStorage.token, prompt).catch((error) => { | |
toast.error(error); | |
return null; | |
}); | |
if (res) { | |
if (res.documents[0].length > 0) { | |
userContext = res.documents[0].reduce((acc, doc, index) => { | |
const createdAtTimestamp = res.metadatas[0][index].created_at; | |
const createdAtDate = new Date(createdAtTimestamp * 1000) | |
.toISOString() | |
.split('T')[0]; | |
return `${acc}${index + 1}. [${createdAtDate}]. ${doc}\n`; | |
}, ''); | |
} | |
console.log(userContext); | |
} | |
} | |
} | |
responseMessage.userContext = userContext; | |
const chatEventEmitter = await getChatEventEmitter(model.id, _chatId); | |
if (webSearchEnabled) { | |
await getWebSearchResults(model.id, parentId, responseMessageId); | |
} | |
let _response = null; | |
if (model?.owned_by === 'openai') { | |
_response = await sendPromptOpenAI(model, prompt, responseMessageId, _chatId); | |
} else if (model) { | |
_response = await sendPromptOllama(model, prompt, responseMessageId, _chatId); | |
} | |
_responses.push(_response); | |
if (chatEventEmitter) clearInterval(chatEventEmitter); | |
} else { | |
toast.error($i18n.t(`Model {{modelId}} not found`, { modelId })); | |
} | |
}) | |
); | |
await chats.set(await getChatList(localStorage.token)); | |
return _responses; | |
}; | |
const sendPromptOllama = async (model, userPrompt, responseMessageId, _chatId) => { | |
let _response = null; | |
const responseMessage = history.messages[responseMessageId]; | |
const userMessage = history.messages[responseMessage.parentId]; | |
// Wait until history/message have been updated | |
await tick(); | |
// Scroll down | |
scrollToBottom(); | |
const messagesBody = [ | |
params?.system || $settings.system || (responseMessage?.userContext ?? null) | |
? { | |
role: 'system', | |
content: `${promptTemplate( | |
params?.system ?? $settings?.system ?? '', | |
$user.name, | |
$settings?.userLocation | |
? await getAndUpdateUserLocation(localStorage.token) | |
: undefined | |
)}${ | |
responseMessage?.userContext ?? null | |
? `\n\nUser Context:\n${responseMessage?.userContext ?? ''}` | |
: '' | |
}` | |
} | |
: undefined, | |
...messages | |
] | |
.filter((message) => message?.content?.trim()) | |
.map((message, idx, arr) => { | |
// Prepare the base message object | |
const baseMessage = { | |
role: message.role, | |
content: message.content | |
}; | |
// Extract and format image URLs if any exist | |
const imageUrls = message.files | |
?.filter((file) => file.type === 'image') | |
.map((file) => file.url.slice(file.url.indexOf(',') + 1)); | |
// Add images array only if it contains elements | |
if (imageUrls && imageUrls.length > 0 && message.role === 'user') { | |
baseMessage.images = imageUrls; | |
} | |
return baseMessage; | |
}); | |
let lastImageIndex = -1; | |
// Find the index of the last object with images | |
messagesBody.forEach((item, index) => { | |
if (item.images) { | |
lastImageIndex = index; | |
} | |
}); | |
// Remove images from all but the last one | |
messagesBody.forEach((item, index) => { | |
if (index !== lastImageIndex) { | |
delete item.images; | |
} | |
}); | |
let files = JSON.parse(JSON.stringify(chatFiles)); | |
if (model?.info?.meta?.knowledge ?? false) { | |
files.push(...model.info.meta.knowledge); | |
} | |
files.push( | |
...(userMessage?.files ?? []).filter((item) => | |
['doc', 'file', 'collection'].includes(item.type) | |
), | |
...(responseMessage?.files ?? []).filter((item) => ['web_search_results'].includes(item.type)) | |
); | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat:start', { | |
detail: { | |
id: responseMessageId | |
} | |
}) | |
); | |
await tick(); | |
const [res, controller] = await generateChatCompletion(localStorage.token, { | |
stream: true, | |
model: model.id, | |
messages: messagesBody, | |
options: { | |
...(params ?? $settings.params ?? {}), | |
stop: | |
params?.stop ?? $settings?.params?.stop ?? undefined | |
? (params?.stop ?? $settings.params.stop).map((str) => | |
decodeURIComponent(JSON.parse('"' + str.replace(/\"/g, '\\"') + '"')) | |
) | |
: undefined, | |
num_predict: params?.max_tokens ?? $settings?.params?.max_tokens ?? undefined, | |
repeat_penalty: | |
params?.frequency_penalty ?? $settings?.params?.frequency_penalty ?? undefined | |
}, | |
format: $settings.requestFormat ?? undefined, | |
keep_alive: $settings.keepAlive ?? undefined, | |
tool_ids: selectedToolIds.length > 0 ? selectedToolIds : undefined, | |
files: files.length > 0 ? files : undefined, | |
...(Object.keys(valves).length ? { valves } : {}), | |
session_id: $socket?.id, | |
chat_id: $chatId, | |
id: responseMessageId | |
}); | |
if (res && res.ok) { | |
console.log('controller', controller); | |
const reader = res.body | |
.pipeThrough(new TextDecoderStream()) | |
.pipeThrough(splitStream('\n')) | |
.getReader(); | |
while (true) { | |
const { value, done } = await reader.read(); | |
if (done || stopResponseFlag || _chatId !== $chatId) { | |
responseMessage.done = true; | |
messages = messages; | |
if (stopResponseFlag) { | |
controller.abort('User: Stop Response'); | |
} else { | |
const messages = createMessagesList(responseMessageId); | |
await chatCompletedHandler(_chatId, model.id, responseMessageId, messages); | |
} | |
_response = responseMessage.content; | |
break; | |
} | |
try { | |
let lines = value.split('\n'); | |
for (const line of lines) { | |
if (line !== '') { | |
console.log(line); | |
let data = JSON.parse(line); | |
if ('citations' in data) { | |
responseMessage.citations = data.citations; | |
continue; | |
} | |
if ('detail' in data) { | |
throw data; | |
} | |
if (data.done == false) { | |
if (responseMessage.content == '' && data.message.content == '\n') { | |
continue; | |
} else { | |
responseMessage.content += data.message.content; | |
const sentences = extractSentencesForAudio(responseMessage.content); | |
sentences.pop(); | |
// dispatch only last sentence and make sure it hasn't been dispatched before | |
if ( | |
sentences.length > 0 && | |
sentences[sentences.length - 1] !== responseMessage.lastSentence | |
) { | |
responseMessage.lastSentence = sentences[sentences.length - 1]; | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat', { | |
detail: { id: responseMessageId, content: sentences[sentences.length - 1] } | |
}) | |
); | |
} | |
messages = messages; | |
} | |
} else { | |
responseMessage.done = true; | |
if (responseMessage.content == '') { | |
responseMessage.error = { | |
code: 400, | |
content: `Oops! No text generated from Ollama, Please try again.` | |
}; | |
} | |
responseMessage.context = data.context ?? null; | |
responseMessage.info = { | |
total_duration: data.total_duration, | |
load_duration: data.load_duration, | |
sample_count: data.sample_count, | |
sample_duration: data.sample_duration, | |
prompt_eval_count: data.prompt_eval_count, | |
prompt_eval_duration: data.prompt_eval_duration, | |
eval_count: data.eval_count, | |
eval_duration: data.eval_duration | |
}; | |
messages = messages; | |
if ($settings.notificationEnabled && !document.hasFocus()) { | |
const notification = new Notification(`${model.id}`, { | |
body: responseMessage.content, | |
icon: `${WEBUI_BASE_URL}/static/favicon.png` | |
}); | |
} | |
if ($settings?.responseAutoCopy ?? false) { | |
copyToClipboard(responseMessage.content); | |
} | |
if ($settings.responseAutoPlayback && !$showCallOverlay) { | |
await tick(); | |
document.getElementById(`speak-button-${responseMessage.id}`)?.click(); | |
} | |
} | |
} | |
} | |
} catch (error) { | |
console.log(error); | |
if ('detail' in error) { | |
toast.error(error.detail); | |
} | |
break; | |
} | |
if (autoScroll) { | |
scrollToBottom(); | |
} | |
} | |
if ($chatId == _chatId) { | |
if ($settings.saveChatHistory ?? true) { | |
chat = await updateChatById(localStorage.token, _chatId, { | |
messages: messages, | |
history: history, | |
models: selectedModels, | |
params: params, | |
files: chatFiles | |
}); | |
await chats.set(await getChatList(localStorage.token)); | |
} | |
} | |
} else { | |
if (res !== null) { | |
const error = await res.json(); | |
console.log(error); | |
if ('detail' in error) { | |
toast.error(error.detail); | |
responseMessage.error = { content: error.detail }; | |
} else { | |
toast.error(error.error); | |
responseMessage.error = { content: error.error }; | |
} | |
} else { | |
toast.error( | |
$i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { provider: 'Ollama' }) | |
); | |
responseMessage.error = { | |
content: $i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { | |
provider: 'Ollama' | |
}) | |
}; | |
} | |
responseMessage.done = true; | |
messages = messages; | |
} | |
stopResponseFlag = false; | |
await tick(); | |
let lastSentence = extractSentencesForAudio(responseMessage.content)?.at(-1) ?? ''; | |
if (lastSentence) { | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat', { | |
detail: { id: responseMessageId, content: lastSentence } | |
}) | |
); | |
} | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat:finish', { | |
detail: { | |
id: responseMessageId, | |
content: responseMessage.content | |
} | |
}) | |
); | |
if (autoScroll) { | |
scrollToBottom(); | |
} | |
if (messages.length == 2 && messages.at(1).content !== '' && selectedModels[0] === model.id) { | |
window.history.replaceState(history.state, '', `/c/${_chatId}`); | |
const _title = await generateChatTitle(userPrompt); | |
await setChatTitle(_chatId, _title); | |
} | |
return _response; | |
}; | |
const sendPromptOpenAI = async (model, userPrompt, responseMessageId, _chatId) => { | |
let _response = null; | |
const responseMessage = history.messages[responseMessageId]; | |
const userMessage = history.messages[responseMessage.parentId]; | |
let files = JSON.parse(JSON.stringify(chatFiles)); | |
if (model?.info?.meta?.knowledge ?? false) { | |
files.push(...model.info.meta.knowledge); | |
} | |
files.push( | |
...(userMessage?.files ?? []).filter((item) => | |
['doc', 'file', 'collection'].includes(item.type) | |
), | |
...(responseMessage?.files ?? []).filter((item) => ['web_search_results'].includes(item.type)) | |
); | |
scrollToBottom(); | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat:start', { | |
detail: { | |
id: responseMessageId | |
} | |
}) | |
); | |
await tick(); | |
try { | |
const [res, controller] = await generateOpenAIChatCompletion( | |
localStorage.token, | |
{ | |
stream: true, | |
model: model.id, | |
stream_options: | |
model.info?.meta?.capabilities?.usage ?? false | |
? { | |
include_usage: true | |
} | |
: undefined, | |
messages: [ | |
params?.system || $settings.system || (responseMessage?.userContext ?? null) | |
? { | |
role: 'system', | |
content: `${promptTemplate( | |
params?.system ?? $settings?.system ?? '', | |
$user.name, | |
$settings?.userLocation | |
? await getAndUpdateUserLocation(localStorage.token) | |
: undefined | |
)}${ | |
responseMessage?.userContext ?? null | |
? `\n\nUser Context:\n${responseMessage?.userContext ?? ''}` | |
: '' | |
}` | |
} | |
: undefined, | |
...messages | |
] | |
.filter((message) => message?.content?.trim()) | |
.map((message, idx, arr) => ({ | |
role: message.role, | |
...((message.files?.filter((file) => file.type === 'image').length > 0 ?? false) && | |
message.role === 'user' | |
? { | |
content: [ | |
{ | |
type: 'text', | |
text: | |
arr.length - 1 !== idx | |
? message.content | |
: message?.raContent ?? message.content | |
}, | |
...message.files | |
.filter((file) => file.type === 'image') | |
.map((file) => ({ | |
type: 'image_url', | |
image_url: { | |
url: file.url | |
} | |
})) | |
] | |
} | |
: { | |
content: | |
arr.length - 1 !== idx | |
? message.content | |
: message?.raContent ?? message.content | |
}) | |
})), | |
seed: params?.seed ?? $settings?.params?.seed ?? undefined, | |
stop: | |
params?.stop ?? $settings?.params?.stop ?? undefined | |
? (params?.stop ?? $settings.params.stop).map((str) => | |
decodeURIComponent(JSON.parse('"' + str.replace(/\"/g, '\\"') + '"')) | |
) | |
: undefined, | |
temperature: params?.temperature ?? $settings?.params?.temperature ?? undefined, | |
top_p: params?.top_p ?? $settings?.params?.top_p ?? undefined, | |
frequency_penalty: | |
params?.frequency_penalty ?? $settings?.params?.frequency_penalty ?? undefined, | |
max_tokens: params?.max_tokens ?? $settings?.params?.max_tokens ?? undefined, | |
tool_ids: selectedToolIds.length > 0 ? selectedToolIds : undefined, | |
files: files.length > 0 ? files : undefined, | |
...(Object.keys(valves).length ? { valves } : {}), | |
session_id: $socket?.id, | |
chat_id: $chatId, | |
id: responseMessageId | |
}, | |
`${WEBUI_BASE_URL}/api` | |
); | |
// Wait until history/message have been updated | |
await tick(); | |
scrollToBottom(); | |
if (res && res.ok && res.body) { | |
const textStream = await createOpenAITextStream(res.body, $settings.splitLargeChunks); | |
let lastUsage = null; | |
for await (const update of textStream) { | |
const { value, done, citations, error, usage } = update; | |
if (error) { | |
await handleOpenAIError(error, null, model, responseMessage); | |
break; | |
} | |
if (done || stopResponseFlag || _chatId !== $chatId) { | |
responseMessage.done = true; | |
messages = messages; | |
if (stopResponseFlag) { | |
controller.abort('User: Stop Response'); | |
} else { | |
const messages = createMessagesList(responseMessageId); | |
await chatCompletedHandler(_chatId, model.id, responseMessageId, messages); | |
} | |
_response = responseMessage.content; | |
break; | |
} | |
if (usage) { | |
lastUsage = usage; | |
} | |
if (citations) { | |
responseMessage.citations = citations; | |
continue; | |
} | |
if (responseMessage.content == '' && value == '\n') { | |
continue; | |
} else { | |
responseMessage.content += value; | |
const sentences = extractSentencesForAudio(responseMessage.content); | |
sentences.pop(); | |
// dispatch only last sentence and make sure it hasn't been dispatched before | |
if ( | |
sentences.length > 0 && | |
sentences[sentences.length - 1] !== responseMessage.lastSentence | |
) { | |
responseMessage.lastSentence = sentences[sentences.length - 1]; | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat', { | |
detail: { id: responseMessageId, content: sentences[sentences.length - 1] } | |
}) | |
); | |
} | |
messages = messages; | |
} | |
if (autoScroll) { | |
scrollToBottom(); | |
} | |
} | |
if ($settings.notificationEnabled && !document.hasFocus()) { | |
const notification = new Notification(`${model.id}`, { | |
body: responseMessage.content, | |
icon: `${WEBUI_BASE_URL}/static/favicon.png` | |
}); | |
} | |
if ($settings.responseAutoCopy) { | |
copyToClipboard(responseMessage.content); | |
} | |
if ($settings.responseAutoPlayback && !$showCallOverlay) { | |
await tick(); | |
document.getElementById(`speak-button-${responseMessage.id}`)?.click(); | |
} | |
if (lastUsage) { | |
responseMessage.info = { ...lastUsage, openai: true }; | |
} | |
if ($chatId == _chatId) { | |
if ($settings.saveChatHistory ?? true) { | |
chat = await updateChatById(localStorage.token, _chatId, { | |
models: selectedModels, | |
messages: messages, | |
history: history, | |
params: params, | |
files: chatFiles | |
}); | |
await chats.set(await getChatList(localStorage.token)); | |
} | |
} | |
} else { | |
await handleOpenAIError(null, res, model, responseMessage); | |
} | |
} catch (error) { | |
await handleOpenAIError(error, null, model, responseMessage); | |
} | |
messages = messages; | |
stopResponseFlag = false; | |
await tick(); | |
let lastSentence = extractSentencesForAudio(responseMessage.content)?.at(-1) ?? ''; | |
if (lastSentence) { | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat', { | |
detail: { id: responseMessageId, content: lastSentence } | |
}) | |
); | |
} | |
eventTarget.dispatchEvent( | |
new CustomEvent('chat:finish', { | |
detail: { | |
id: responseMessageId, | |
content: responseMessage.content | |
} | |
}) | |
); | |
if (autoScroll) { | |
scrollToBottom(); | |
} | |
if (messages.length == 2 && selectedModels[0] === model.id) { | |
window.history.replaceState(history.state, '', `/c/${_chatId}`); | |
const _title = await generateChatTitle(userPrompt); | |
await setChatTitle(_chatId, _title); | |
} | |
return _response; | |
}; | |
const handleOpenAIError = async (error, res: Response | null, model, responseMessage) => { | |
let errorMessage = ''; | |
let innerError; | |
if (error) { | |
innerError = error; | |
} else if (res !== null) { | |
innerError = await res.json(); | |
} | |
console.error(innerError); | |
if ('detail' in innerError) { | |
toast.error(innerError.detail); | |
errorMessage = innerError.detail; | |
} else if ('error' in innerError) { | |
if ('message' in innerError.error) { | |
toast.error(innerError.error.message); | |
errorMessage = innerError.error.message; | |
} else { | |
toast.error(innerError.error); | |
errorMessage = innerError.error; | |
} | |
} else if ('message' in innerError) { | |
toast.error(innerError.message); | |
errorMessage = innerError.message; | |
} | |
responseMessage.error = { | |
content: | |
$i18n.t(`Uh-oh! There was an issue connecting to {{provider}}.`, { | |
provider: model.name ?? model.id | |
}) + | |
'\n' + | |
errorMessage | |
}; | |
responseMessage.done = true; | |
messages = messages; | |
}; | |
const stopResponse = () => { | |
stopResponseFlag = true; | |
console.log('stopResponse'); | |
}; | |
const regenerateResponse = async (message) => { | |
console.log('regenerateResponse'); | |
if (messages.length != 0) { | |
let userMessage = history.messages[message.parentId]; | |
let userPrompt = userMessage.content; | |
if ((userMessage?.models ?? [...selectedModels]).length == 1) { | |
// If user message has only one model selected, sendPrompt automatically selects it for regeneration | |
await sendPrompt(userPrompt, userMessage.id); | |
} else { | |
// If there are multiple models selected, use the model of the response message for regeneration | |
// e.g. many model chat | |
await sendPrompt(userPrompt, userMessage.id, { modelId: message.model }); | |
} | |
} | |
}; | |
const continueGeneration = async () => { | |
console.log('continueGeneration'); | |
const _chatId = JSON.parse(JSON.stringify($chatId)); | |
if (messages.length != 0 && messages.at(-1).done == true) { | |
const responseMessage = history.messages[history.currentId]; | |
responseMessage.done = false; | |
await tick(); | |
const model = $models.filter((m) => m.id === responseMessage.model).at(0); | |
if (model) { | |
if (model?.owned_by === 'openai') { | |
await sendPromptOpenAI( | |
model, | |
history.messages[responseMessage.parentId].content, | |
responseMessage.id, | |
_chatId | |
); | |
} else | |
await sendPromptOllama( | |
model, | |
history.messages[responseMessage.parentId].content, | |
responseMessage.id, | |
_chatId | |
); | |
} | |
} else { | |
toast.error($i18n.t(`Model {{modelId}} not found`, { modelId })); | |
} | |
}; | |
const generateChatTitle = async (userPrompt) => { | |
if ($settings?.title?.auto ?? true) { | |
const title = await generateTitle( | |
localStorage.token, | |
selectedModels[0], | |
userPrompt, | |
$chatId | |
).catch((error) => { | |
console.error(error); | |
return 'New Chat'; | |
}); | |
return title; | |
} else { | |
return `${userPrompt}`; | |
} | |
}; | |
const setChatTitle = async (_chatId, _title) => { | |
if (_chatId === $chatId) { | |
title = _title; | |
} | |
if ($settings.saveChatHistory ?? true) { | |
chat = await updateChatById(localStorage.token, _chatId, { title: _title }); | |
await chats.set(await getChatList(localStorage.token)); | |
} | |
}; | |
const getWebSearchResults = async (model: string, parentId: string, responseId: string) => { | |
const responseMessage = history.messages[responseId]; | |
const userMessage = history.messages[parentId]; | |
responseMessage.statusHistory = [ | |
{ | |
done: false, | |
action: 'web_search', | |
description: $i18n.t('Generating search query') | |
} | |
]; | |
messages = messages; | |
const prompt = userMessage.content; | |
let searchQuery = await generateSearchQuery(localStorage.token, model, messages, prompt).catch( | |
(error) => { | |
console.log(error); | |
return prompt; | |
} | |
); | |
if (!searchQuery) { | |
toast.warning($i18n.t('No search query generated')); | |
responseMessage.statusHistory.push({ | |
done: true, | |
error: true, | |
action: 'web_search', | |
description: 'No search query generated' | |
}); | |
messages = messages; | |
} | |
responseMessage.statusHistory.push({ | |
done: false, | |
action: 'web_search', | |
description: $i18n.t(`Searching "{{searchQuery}}"`, { searchQuery }) | |
}); | |
messages = messages; | |
const results = await runWebSearch(localStorage.token, searchQuery).catch((error) => { | |
console.log(error); | |
toast.error(error); | |
return null; | |
}); | |
if (results) { | |
responseMessage.statusHistory.push({ | |
done: true, | |
action: 'web_search', | |
description: $i18n.t('Searched {{count}} sites', { count: results.filenames.length }), | |
query: searchQuery, | |
urls: results.filenames | |
}); | |
if (responseMessage?.files ?? undefined === undefined) { | |
responseMessage.files = []; | |
} | |
responseMessage.files.push({ | |
collection_name: results.collection_name, | |
name: searchQuery, | |
type: 'web_search_results', | |
urls: results.filenames | |
}); | |
messages = messages; | |
} else { | |
responseMessage.statusHistory.push({ | |
done: true, | |
error: true, | |
action: 'web_search', | |
description: 'No search results found' | |
}); | |
messages = messages; | |
} | |
}; | |
const getTags = async () => { | |
return await getTagsById(localStorage.token, $chatId).catch(async (error) => { | |
return []; | |
}); | |
}; | |
</script> | |
<svelte:head> | |
<title> | |
{title | |
? `${title.length > 30 ? `${title.slice(0, 30)}...` : title} | ${$WEBUI_NAME}` | |
: `${$WEBUI_NAME}`} | |
</title> | |
</svelte:head> | |
<audio id="audioElement" src="" style="display: none;" /> | |
<EventConfirmDialog | |
bind:show={showEventConfirmation} | |
title={eventConfirmationTitle} | |
message={eventConfirmationMessage} | |
input={eventConfirmationInput} | |
inputPlaceholder={eventConfirmationInputPlaceholder} | |
on:confirm={(e) => { | |
if (e.detail) { | |
eventCallback(e.detail); | |
} else { | |
eventCallback(true); | |
} | |
}} | |
on:cancel={() => { | |
eventCallback(false); | |
}} | |
/> | |
{#if $showCallOverlay} | |
<CallOverlay | |
{submitPrompt} | |
{stopResponse} | |
bind:files | |
modelId={selectedModelIds?.at(0) ?? null} | |
chatId={$chatId} | |
{eventTarget} | |
/> | |
{/if} | |
{#if !chatIdProp || (loaded && chatIdProp)} | |
<div | |
class="h-screen max-h-[100dvh] {$showSidebar | |
? 'md:max-w-[calc(100%-260px)]' | |
: ''} w-full max-w-full flex flex-col" | |
> | |
{#if $settings?.backgroundImageUrl ?? null} | |
<div | |
class="absolute {$showSidebar | |
? 'md:max-w-[calc(100%-260px)] md:translate-x-[260px]' | |
: ''} top-0 left-0 w-full h-full bg-cover bg-center bg-no-repeat" | |
style="background-image: url({$settings.backgroundImageUrl}) " | |
/> | |
<div | |
class="absolute top-0 left-0 w-full h-full bg-gradient-to-t from-white to-white/85 dark:from-gray-900 dark:to-[#171717]/90 z-0" | |
/> | |
{/if} | |
<Navbar | |
{title} | |
bind:selectedModels | |
bind:showModelSelector | |
bind:showControls | |
shareEnabled={messages.length > 0} | |
{chat} | |
{initNewChat} | |
/> | |
{#if $banners.length > 0 && messages.length === 0 && !$chatId && selectedModels.length <= 1} | |
<div | |
class="absolute top-[4.25rem] w-full {$showSidebar | |
? 'md:max-w-[calc(100%-260px)]' | |
: ''} {showControls ? 'lg:pr-[24rem]' : ''} z-20" | |
> | |
<div class=" flex flex-col gap-1 w-full"> | |
{#each $banners.filter( (b) => (b.dismissible ? !JSON.parse(localStorage.getItem('dismissedBannerIds') ?? '[]').includes(b.id) : true) ) as banner} | |
<Banner | |
{banner} | |
on:dismiss={(e) => { | |
const bannerId = e.detail; | |
localStorage.setItem( | |
'dismissedBannerIds', | |
JSON.stringify( | |
[ | |
bannerId, | |
...JSON.parse(localStorage.getItem('dismissedBannerIds') ?? '[]') | |
].filter((id) => $banners.find((b) => b.id === id)) | |
) | |
); | |
}} | |
/> | |
{/each} | |
</div> | |
</div> | |
{/if} | |
<div class="flex flex-col flex-auto z-10"> | |
<div | |
class=" pb-2.5 flex flex-col justify-between w-full flex-auto overflow-auto h-0 max-w-full z-10 scrollbar-hidden {showControls | |
? 'lg:pr-[24rem]' | |
: ''}" | |
id="messages-container" | |
bind:this={messagesContainerElement} | |
on:scroll={(e) => { | |
autoScroll = | |
messagesContainerElement.scrollHeight - messagesContainerElement.scrollTop <= | |
messagesContainerElement.clientHeight + 5; | |
}} | |
> | |
<div class=" h-full w-full flex flex-col {chatIdProp ? 'py-4' : 'pt-2 pb-4'}"> | |
<Messages | |
chatId={$chatId} | |
{selectedModels} | |
{processing} | |
bind:history | |
bind:messages | |
bind:autoScroll | |
bind:prompt | |
bottomPadding={files.length > 0} | |
{sendPrompt} | |
{continueGeneration} | |
{regenerateResponse} | |
{chatActionHandler} | |
/> | |
</div> | |
</div> | |
<div class={showControls ? 'lg:pr-[24rem]' : ''}> | |
<MessageInput | |
bind:files | |
bind:prompt | |
bind:autoScroll | |
bind:selectedToolIds | |
bind:webSearchEnabled | |
bind:atSelectedModel | |
availableToolIds={selectedModelIds.reduce((a, e, i, arr) => { | |
const model = $models.find((m) => m.id === e); | |
if (model?.info?.meta?.toolIds ?? false) { | |
return [...new Set([...a, ...model.info.meta.toolIds])]; | |
} | |
return a; | |
}, [])} | |
transparentBackground={$settings?.backgroundImageUrl ?? false} | |
{selectedModels} | |
{messages} | |
{submitPrompt} | |
{stopResponse} | |
/> | |
</div> | |
</div> | |
<ChatControls | |
models={selectedModelIds.reduce((a, e, i, arr) => { | |
const model = $models.find((m) => m.id === e); | |
if (model) { | |
return [...a, model]; | |
} | |
return a; | |
}, [])} | |
bind:show={showControls} | |
bind:chatFiles | |
bind:params | |
bind:valves | |
/> | |
</div> | |
{/if} | |