/* https://github.com/david-haerer/chatapi MIT License Copyright (c) 2023 David Härer Copyright (c) 2024 Ettore Di Giacinto Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ // Track requests per chat ID to support parallel chatting let activeRequests = new Map(); // chatId -> { controller, reader, startTime, tokensReceived, interval, maxTokensPerSecond } // Global variables for UI (stop button, etc.) let currentAbortController = null; // For stop button - tracks the active chat's request let currentReader = null; let tokensPerSecondInterval = null; let tokensPerSecondIntervalChatId = null; // Track which chat the interval is for let lastTokensPerSecond = null; // Store the last calculated rate // Storage key for chats const CHATS_STORAGE_KEY = 'localai_chats_data'; const SYSTEM_PROMPT_STORAGE_KEY = 'system_prompt'; // Old key for migration // Debounce timer for auto-save let saveDebounceTimer = null; const SAVE_DEBOUNCE_MS = 500; // Save chats to localStorage with error handling function saveChatsToStorage() { if (!window.Alpine || !Alpine.store("chat")) { return false; } try { const chatStore = Alpine.store("chat"); const data = { chats: chatStore.chats.map(chat => ({ id: chat.id, name: chat.name, model: chat.model, history: chat.history, systemPrompt: chat.systemPrompt, mcpMode: chat.mcpMode, tokenUsage: chat.tokenUsage, contextSize: chat.contextSize, createdAt: chat.createdAt, updatedAt: chat.updatedAt })), activeChatId: chatStore.activeChatId, lastSaved: Date.now() }; const jsonData = JSON.stringify(data); localStorage.setItem(CHATS_STORAGE_KEY, jsonData); return true; } catch (error) { // Handle quota exceeded or other storage errors if (error.name === 'QuotaExceededError' || error.code === 22) { console.warn('localStorage quota exceeded. Consider cleaning up old chats.'); // Try to save without history (last resort) try { const chatStore = Alpine.store("chat"); const data = { chats: chatStore.chats.map(chat => ({ id: chat.id, name: chat.name, model: chat.model, history: [], // Clear history to save space systemPrompt: chat.systemPrompt, mcpMode: chat.mcpMode, tokenUsage: chat.tokenUsage, contextSize: chat.contextSize, createdAt: chat.createdAt, updatedAt: chat.updatedAt })), activeChatId: chatStore.activeChatId, lastSaved: Date.now() }; localStorage.setItem(CHATS_STORAGE_KEY, JSON.stringify(data)); return true; } catch (e2) { console.error('Failed to save chats even without history:', e2); return false; } } else { console.error('Error saving chats to localStorage:', error); return false; } } } // Load chats from localStorage with migration support function loadChatsFromStorage() { try { const stored = localStorage.getItem(CHATS_STORAGE_KEY); if (stored) { const data = JSON.parse(stored); // Validate structure if (data && Array.isArray(data.chats)) { return { chats: data.chats, activeChatId: data.activeChatId || null, lastSaved: data.lastSaved || null }; } } // Migration: Check for old format const oldSystemPrompt = localStorage.getItem(SYSTEM_PROMPT_STORAGE_KEY); if (oldSystemPrompt) { // Migrate old single-chat format to new multi-chat format const chatStore = Alpine.store("chat"); if (chatStore) { const migratedChat = chatStore.createChat( document.getElementById("chat-model")?.value || "", oldSystemPrompt, false ); // Try to preserve any existing history if available if (chatStore.activeChat()) { chatStore.activeChat().name = "Migrated Chat"; } // Save migrated data saveChatsToStorage(); // Remove old key localStorage.removeItem(SYSTEM_PROMPT_STORAGE_KEY); return { chats: chatStore.chats, activeChatId: chatStore.activeChatId, lastSaved: Date.now() }; } } return null; } catch (error) { console.error('Error loading chats from localStorage:', error); // Try to recover by clearing corrupted data try { localStorage.removeItem(CHATS_STORAGE_KEY); } catch (e) { console.error('Failed to clear corrupted data:', e); } return null; } } // Auto-save with debouncing function autoSaveChats() { if (saveDebounceTimer) { clearTimeout(saveDebounceTimer); } saveDebounceTimer = setTimeout(() => { saveChatsToStorage(); }, SAVE_DEBOUNCE_MS); } // Function to check if a chat has an active request (for UI indicators) function isChatRequestActive(chatId) { if (!chatId || !activeRequests) { return false; } const request = activeRequests.get(chatId); return request && (request.controller || request.reader); } // Helper function to update reactive tracking for UI indicators function updateRequestTracking(chatId, isActive) { const chatStore = Alpine.store("chat"); if (chatStore && typeof chatStore.updateActiveRequestTracking === 'function') { chatStore.updateActiveRequestTracking(chatId, isActive); } } // Make functions available globally window.autoSaveChats = autoSaveChats; window.createNewChat = createNewChat; window.switchChat = switchChat; window.deleteChat = deleteChat; window.bulkDeleteChats = bulkDeleteChats; window.updateChatName = updateChatName; window.updateUIForActiveChat = updateUIForActiveChat; window.isChatRequestActive = isChatRequestActive; // Create a new chat function createNewChat(model, systemPrompt, mcpMode) { if (!window.Alpine || !Alpine.store("chat")) { return null; } const chatStore = Alpine.store("chat"); const chat = chatStore.createChat(model, systemPrompt, mcpMode); // Save to storage saveChatsToStorage(); // Update UI to reflect new active chat updateUIForActiveChat(); return chat; } // Switch to a different chat function switchChat(chatId) { if (!window.Alpine || !Alpine.store("chat")) { return false; } const chatStore = Alpine.store("chat"); const oldActiveChat = chatStore.activeChat(); if (chatStore.switchChat(chatId)) { // CRITICAL: Stop interval FIRST before any other operations // This prevents the interval from updating with wrong chat's data if (tokensPerSecondInterval) { clearInterval(tokensPerSecondInterval); tokensPerSecondInterval = null; } // Immediately clear the display to prevent showing stale data const tokensPerSecondDisplay = document.getElementById('tokens-per-second'); if (tokensPerSecondDisplay) { tokensPerSecondDisplay.textContent = '-'; } // Save current state before switching saveChatsToStorage(); // Hide badge when switching chats - will be shown if new chat has completed request const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (maxBadge) { maxBadge.style.display = 'none'; } // Update global request tracking for stop button (only if new chat has active request) const newActiveChat = chatStore.activeChat(); const newRequest = activeRequests.get(newActiveChat?.id); if (newRequest) { currentAbortController = newRequest.controller; currentReader = newRequest.reader; // Update loader state if new chat has active request const hasActiveRequest = newRequest.controller || newRequest.reader; if (hasActiveRequest) { toggleLoader(true, newActiveChat.id); // Wait a bit to ensure switch is complete and interval is stopped setTimeout(() => { // Double-check we're still on the same chat and interval is stopped const currentActiveChat = chatStore.activeChat(); if (currentActiveChat && currentActiveChat.id === newActiveChat.id) { // Make absolutely sure interval is stopped if (tokensPerSecondInterval) { clearInterval(tokensPerSecondInterval); tokensPerSecondInterval = null; tokensPerSecondIntervalChatId = null; } // Update display for the new active chat updateTokensPerSecond(newActiveChat.id); // Restart interval to pick up the new active chat startTokensPerSecondInterval(); } }, 100); } else { toggleLoader(false, newActiveChat.id); } } else { // No active request for new chat, clear global references currentAbortController = null; currentReader = null; toggleLoader(false, newActiveChat?.id); // Display is already cleared above // Check if this chat has a completed request with max tokens/s to show // Note: We only show badge for completed requests, not active ones // The badge will be shown when the request ends, not when switching to a chat } // Update UI to reflect new active chat updateUIForActiveChat(); return true; } return false; } // Delete a chat function deleteChat(chatId) { if (!window.Alpine || !Alpine.store("chat")) { return false; } const chatStore = Alpine.store("chat"); // Prevent deleting the last chat if (chatStore.chats.length <= 1) { alert('Cannot delete the last chat. Please create a new chat first.'); return false; } if (chatStore.deleteChat(chatId)) { // Ensure at least one chat exists after deletion if (chatStore.chats.length === 0) { const currentModel = document.getElementById("chat-model")?.value || ""; chatStore.createChat(currentModel, "", false); } saveChatsToStorage(); updateUIForActiveChat(); return true; } return false; } // Bulk delete chats function bulkDeleteChats(options) { if (!window.Alpine || !Alpine.store("chat")) { return 0; } const chatStore = Alpine.store("chat"); let deletedCount = 0; const now = Date.now(); if (options.deleteAll) { // Delete all chats except active one, or create new if deleting all const activeId = chatStore.activeChatId; chatStore.chats = chatStore.chats.filter(chat => { if (chat.id === activeId && chatStore.chats.length > 1) { return true; // Keep active chat if there are others } deletedCount++; return false; }); // If all deleted, create a new chat if (chatStore.chats.length === 0) { chatStore.createChat(); } else if (!chatStore.chats.find(c => c.id === activeId)) { // Active chat was deleted, switch to first available if (chatStore.chats.length > 0) { chatStore.activeChatId = chatStore.chats[0].id; } } } else if (options.olderThanDays) { const cutoffTime = now - (options.olderThanDays * 24 * 60 * 60 * 1000); const activeId = chatStore.activeChatId; chatStore.chats = chatStore.chats.filter(chat => { if (chat.id === activeId) { return true; // Never delete active chat } if (chat.updatedAt < cutoffTime) { deletedCount++; return false; } return true; }); // Ensure at least one chat exists if (chatStore.chats.length === 0) { const currentModel = document.getElementById("chat-model")?.value || ""; chatStore.createChat(currentModel, "", false); } } if (deletedCount > 0) { saveChatsToStorage(); updateUIForActiveChat(); } return deletedCount; } // Update UI elements to reflect active chat function updateUIForActiveChat() { if (!window.Alpine || !Alpine.store("chat")) { return; } const chatStore = Alpine.store("chat"); // Ensure at least one chat exists if (!chatStore.chats || chatStore.chats.length === 0) { const currentModel = document.getElementById("chat-model")?.value || ""; chatStore.createChat(currentModel, "", false); } const activeChat = chatStore.activeChat(); if (!activeChat) { // No active chat, set first one as active if (chatStore.chats.length > 0) { chatStore.activeChatId = chatStore.chats[0].id; } else { // Still no chats, create one const currentModel = document.getElementById("chat-model")?.value || ""; chatStore.createChat(currentModel, "", false); } return; } // Update system prompt input const systemPromptInput = document.getElementById("systemPrompt"); if (systemPromptInput) { systemPromptInput.value = activeChat.systemPrompt || ""; } // Update MCP toggle const mcpToggle = document.getElementById("mcp-toggle"); if (mcpToggle) { mcpToggle.checked = activeChat.mcpMode || false; } // Update model selector (if needed) const modelSelector = document.getElementById("modelSelector"); if (modelSelector && activeChat.model) { // Find and select the option matching the active chat's model for (let option of modelSelector.options) { if (option.value === `chat/${activeChat.model}` || option.text === activeChat.model) { option.selected = true; break; } } } // Update chat model hidden input const chatModelInput = document.getElementById("chat-model"); if (chatModelInput) { chatModelInput.value = activeChat.model || ""; } } // Update chat name function updateChatName(chatId, name) { if (!window.Alpine || !Alpine.store("chat")) { return false; } const chatStore = Alpine.store("chat"); if (chatStore.updateChatName(chatId, name)) { autoSaveChats(); return true; } return false; } function toggleLoader(show, chatId = null) { const sendButton = document.getElementById('send-button'); const stopButton = document.getElementById('stop-button'); const headerLoadingIndicator = document.getElementById('header-loading-indicator'); const tokensPerSecondDisplay = document.getElementById('tokens-per-second'); if (show) { sendButton.style.display = 'none'; stopButton.style.display = 'block'; if (headerLoadingIndicator) headerLoadingIndicator.style.display = 'block'; // Start updating tokens/second display only if this is for the active chat const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); // Always stop any existing interval first if (tokensPerSecondInterval) { clearInterval(tokensPerSecondInterval); tokensPerSecondInterval = null; } // Use provided chatId or get from active chat const targetChatId = chatId || (activeChat ? activeChat.id : null); if (tokensPerSecondDisplay && targetChatId && activeChat && activeChat.id === targetChatId) { tokensPerSecondDisplay.textContent = '-'; // Hide max badge when starting new request const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (maxBadge) { maxBadge.style.display = 'none'; } // Don't start interval here - it will be started when the request is created // Just update once to show initial state updateTokensPerSecond(targetChatId); } else if (tokensPerSecondDisplay) { // Not the active chat, hide or show dash tokensPerSecondDisplay.textContent = '-'; } } else { sendButton.style.display = 'block'; stopButton.style.display = 'none'; if (headerLoadingIndicator) headerLoadingIndicator.style.display = 'none'; // Stop updating but keep the last value visible only if this was the active chat const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (chatId && activeChat && activeChat.id === chatId) { // Stop the interval since this request is done stopTokensPerSecondInterval(); // Keep the last calculated rate visible if (tokensPerSecondDisplay && lastTokensPerSecond !== null) { tokensPerSecondDisplay.textContent = lastTokensPerSecond; } // Check if there are other active requests for the active chat and restart interval if needed const activeRequest = activeRequests.get(activeChat.id); if (activeRequest && (activeRequest.controller || activeRequest.reader)) { // Restart interval for the active chat startTokensPerSecondInterval(); } } else if (tokensPerSecondDisplay) { // Not the active chat, just show dash tokensPerSecondDisplay.textContent = '-'; } // Only clear global references if this was the active chat if (chatId && activeChat && activeChat.id === chatId) { currentAbortController = null; currentReader = null; // Show the max tokens/s badge when request ends const request = activeRequests.get(chatId); if (request && request.maxTokensPerSecond > 0) { updateMaxTokensPerSecondBadge(chatId, request.maxTokensPerSecond); } } } } // Start a single global interval that updates tokens/second for the active chat function startTokensPerSecondInterval() { // Stop any existing interval first stopTokensPerSecondInterval(); // Get the current active chat ID to track const chatStore = Alpine.store("chat"); if (!chatStore) { return; } const activeChat = chatStore.activeChat(); if (!activeChat) { return; } // Check if active chat has an active request // We can start the interval if we have at least a controller (reader will be set when streaming starts) const request = activeRequests.get(activeChat.id); if (!request) { // No active request for this chat return; } if (!request.controller) { // No controller yet, don't start interval return; } // Store which chat this interval is for tokensPerSecondIntervalChatId = activeChat.id; // Start a single interval that always checks the current active chat // Use a function that always gets fresh state, no closures tokensPerSecondInterval = setInterval(() => { // Always get fresh references - no closures const currentChatStore = Alpine.store("chat"); if (!currentChatStore) { stopTokensPerSecondInterval(); return; } const currentActiveChat = currentChatStore.activeChat(); const tokensPerSecondDisplay = document.getElementById('tokens-per-second'); if (!tokensPerSecondDisplay) { stopTokensPerSecondInterval(); return; } // CRITICAL: Check if the active chat has changed if (!currentActiveChat || currentActiveChat.id !== tokensPerSecondIntervalChatId) { // Active chat changed, stop this interval immediately and hide badge const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (maxBadge) { maxBadge.style.display = 'none'; } stopTokensPerSecondInterval(); return; } // Check if active chat still has an active request const currentRequest = activeRequests.get(currentActiveChat.id); if (!currentRequest) { // No active request for this chat anymore - hide badge tokensPerSecondDisplay.textContent = '-'; const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (maxBadge) { maxBadge.style.display = 'none'; } stopTokensPerSecondInterval(); return; } // If controller is gone, request ended - show max rate badge only for this chat if (!currentRequest.controller) { tokensPerSecondDisplay.textContent = '-'; if (currentRequest.maxTokensPerSecond > 0) { // Only show badge if this is still the active chat updateMaxTokensPerSecondBadge(currentActiveChat.id, currentRequest.maxTokensPerSecond); } else { // Hide badge if no max value const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (maxBadge) { maxBadge.style.display = 'none'; } } stopTokensPerSecondInterval(); return; } // Update for the current active chat only updateTokensPerSecond(currentActiveChat.id); }, 250); // Update more frequently for better responsiveness } // Stop the tokens/second interval function stopTokensPerSecondInterval() { if (tokensPerSecondInterval) { clearInterval(tokensPerSecondInterval); tokensPerSecondInterval = null; } tokensPerSecondIntervalChatId = null; // Clear tracked chat ID const tokensPerSecondDisplay = document.getElementById('tokens-per-second'); if (tokensPerSecondDisplay) { tokensPerSecondDisplay.textContent = '-'; } // Clear the last rate so it doesn't get reused lastTokensPerSecond = null; } function updateTokensPerSecond(chatId) { const tokensPerSecondDisplay = document.getElementById('tokens-per-second'); if (!tokensPerSecondDisplay || !chatId) { return; } // Get the request info for this chat const request = activeRequests.get(chatId); if (!request || !request.startTime) { tokensPerSecondDisplay.textContent = '-'; return; } // Verify the request is still active (controller is cleared when request ends) if (!request.controller) { tokensPerSecondDisplay.textContent = '-'; return; } // Check if this is still the active chat const chatStore = Alpine.store("chat"); const activeChat = chatStore ? chatStore.activeChat() : null; if (!activeChat || activeChat.id !== chatId) { // Not the active chat anymore tokensPerSecondDisplay.textContent = '-'; return; } const elapsedSeconds = (Date.now() - request.startTime) / 1000; // Show rate if we have tokens, otherwise show waiting indicator if (elapsedSeconds > 0) { if (request.tokensReceived > 0) { const rate = request.tokensReceived / elapsedSeconds; // Update max rate if this is higher if (rate > (request.maxTokensPerSecond || 0)) { request.maxTokensPerSecond = rate; } const formattedRate = `${rate.toFixed(1)} tokens/s`; tokensPerSecondDisplay.textContent = formattedRate; lastTokensPerSecond = formattedRate; // Store the last calculated rate // Update the max badge if it exists (only show during active request if user wants, or we can show it at the end) } else { // Request is active but no tokens yet - show waiting tokensPerSecondDisplay.textContent = '0.0 tokens/s'; } } else { // Just started tokensPerSecondDisplay.textContent = '-'; } } // Update the max tokens/s badge display function updateMaxTokensPerSecondBadge(chatId, maxRate) { const maxBadge = document.getElementById('max-tokens-per-second-badge'); if (!maxBadge) return; // Check if this is still the active chat const chatStore = Alpine.store("chat"); const activeChat = chatStore ? chatStore.activeChat() : null; if (!activeChat || activeChat.id !== chatId) { // Not the active chat, hide badge maxBadge.style.display = 'none'; return; } // Only show badge if we have a valid max rate if (maxRate > 0) { maxBadge.textContent = `Peak: ${maxRate.toFixed(1)} tokens/s`; maxBadge.style.display = 'inline-flex'; } else { maxBadge.style.display = 'none'; } } function scrollThinkingBoxToBottom() { // Find all thinking/reasoning message containers that are expanded const thinkingBoxes = document.querySelectorAll('[data-thinking-box]'); thinkingBoxes.forEach(box => { // Only scroll if the box is visible (expanded) and has overflow if (box.offsetParent !== null && box.scrollHeight > box.clientHeight) { box.scrollTo({ top: box.scrollHeight, behavior: 'smooth' }); } }); } // Make function available globally window.scrollThinkingBoxToBottom = scrollThinkingBoxToBottom; function stopRequest() { // Stop the request for the currently active chat const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (!activeChat) return; const request = activeRequests.get(activeChat.id); if (request) { if (request.controller) { request.controller.abort(); } if (request.reader) { request.reader.cancel(); } if (request.interval) { clearInterval(request.interval); } activeRequests.delete(activeChat.id); updateRequestTracking(activeChat.id, false); } // Also clear global references if (currentAbortController) { currentAbortController.abort(); currentAbortController = null; } if (currentReader) { currentReader.cancel(); currentReader = null; } toggleLoader(false, activeChat.id); chatStore.add( "assistant", `Request cancelled by user`, null, null, activeChat.id ); } function processThinkingTags(content) { const thinkingRegex = /(.*?)<\/thinking>|(.*?)<\/think>/gs; const parts = content.split(thinkingRegex); let regularContent = ""; let thinkingContent = ""; for (let i = 0; i < parts.length; i++) { if (i % 3 === 0) { // Regular content regularContent += parts[i]; } else if (i % 3 === 1) { // content thinkingContent = parts[i]; } else if (i % 3 === 2) { // content thinkingContent = parts[i]; } } return { regularContent: regularContent.trim(), thinkingContent: thinkingContent.trim() }; } function submitSystemPrompt(event) { event.preventDefault(); const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (activeChat) { activeChat.systemPrompt = document.getElementById("systemPrompt").value; activeChat.updatedAt = Date.now(); autoSaveChats(); } document.getElementById("systemPrompt").blur(); } function handleShutdownResponse(event, modelName) { // Check if the request was successful if (event.detail.successful) { // Show a success message (optional) console.log(`Model ${modelName} stopped successfully`); // Refresh the page to update the UI window.location.reload(); } else { // Show an error message (optional) console.error(`Failed to stop model ${modelName}`); // You could also show a user-friendly error message here // For now, we'll still refresh to show the current state window.location.reload(); } } var images = []; var audios = []; var fileContents = []; var currentFileNames = []; // Track file names to data URLs for proper removal var imageFileMap = new Map(); // fileName -> dataURL var audioFileMap = new Map(); // fileName -> dataURL async function extractTextFromPDF(pdfData) { try { const pdf = await pdfjsLib.getDocument({ data: pdfData }).promise; let fullText = ''; for (let i = 1; i <= pdf.numPages; i++) { const page = await pdf.getPage(i); const textContent = await page.getTextContent(); const pageText = textContent.items.map(item => item.str).join(' '); fullText += pageText + '\n'; } return fullText; } catch (error) { console.error('Error extracting text from PDF:', error); throw error; } } // Global function to handle file selection and update Alpine.js state window.handleFileSelection = function(event, fileType) { if (!event.target.files || !event.target.files.length) return; // Get the Alpine.js component - find the parent div with x-data containing attachedFiles let inputContainer = event.target.closest('[x-data*="attachedFiles"]'); if (!inputContainer && window.Alpine) { // Fallback: find any element with attachedFiles in x-data inputContainer = document.querySelector('[x-data*="attachedFiles"]'); } if (!inputContainer || !window.Alpine) return; const alpineData = Alpine.$data(inputContainer); if (!alpineData || !alpineData.attachedFiles) return; Array.from(event.target.files).forEach(file => { // Check if file already exists const exists = alpineData.attachedFiles.some(f => f.name === file.name && f.type === fileType); if (!exists) { alpineData.attachedFiles.push({ name: file.name, type: fileType }); // Process the file based on type if (fileType === 'image') { readInputImageFile(file); } else if (fileType === 'audio') { readInputAudioFile(file); } else if (fileType === 'file') { readInputFileFile(file); } } }); }; // Global function to remove file from input window.removeFileFromInput = function(fileType, fileName) { // Remove from arrays if (fileType === 'image') { // Remove from images array using the mapping const dataURL = imageFileMap.get(fileName); if (dataURL) { const imageIndex = images.indexOf(dataURL); if (imageIndex !== -1) { images.splice(imageIndex, 1); } imageFileMap.delete(fileName); } } else if (fileType === 'audio') { // Remove from audios array using the mapping const dataURL = audioFileMap.get(fileName); if (dataURL) { const audioIndex = audios.indexOf(dataURL); if (audioIndex !== -1) { audios.splice(audioIndex, 1); } audioFileMap.delete(fileName); } } else if (fileType === 'file') { // Remove from fileContents and currentFileNames const fileIndex = currentFileNames.indexOf(fileName); if (fileIndex !== -1) { currentFileNames.splice(fileIndex, 1); fileContents.splice(fileIndex, 1); } } // Also remove from the actual input element const inputId = fileType === 'image' ? 'input_image' : fileType === 'audio' ? 'input_audio' : 'input_file'; const input = document.getElementById(inputId); if (input && input.files) { const dt = new DataTransfer(); Array.from(input.files).forEach(file => { if (file.name !== fileName) { dt.items.add(file); } }); input.files = dt.files; } }; function readInputFile() { if (!this.files || !this.files.length) return; Array.from(this.files).forEach(file => { readInputFileFile(file); }); } function readInputFileFile(file) { const FR = new FileReader(); currentFileNames.push(file.name); const fileExtension = file.name.split('.').pop().toLowerCase(); FR.addEventListener("load", async function(evt) { if (fileExtension === 'pdf') { try { const content = await extractTextFromPDF(evt.target.result); fileContents.push({ name: file.name, content: content }); } catch (error) { console.error('Error processing PDF:', error); fileContents.push({ name: file.name, content: "Error processing PDF file" }); } } else { // For text and markdown files fileContents.push({ name: file.name, content: evt.target.result }); } }); if (fileExtension === 'pdf') { FR.readAsArrayBuffer(file); } else { FR.readAsText(file); } } function submitPrompt(event) { event.preventDefault(); const input = document.getElementById("input"); if (!input) return; const inputValue = input.value; if (!inputValue.trim()) return; // Don't send empty messages // Check if there's an active request for the current chat const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (activeChat) { const activeRequest = activeRequests.get(activeChat.id); if (activeRequest && (activeRequest.controller || activeRequest.reader)) { // Abort current request for this chat stopRequest(); // Small delay to ensure cleanup completes setTimeout(() => { // Continue with new request processAndSendMessage(inputValue); }, 100); return; } } processAndSendMessage(inputValue); } function processAndSendMessage(inputValue) { let fullInput = inputValue; // If there are file contents, append them to the input for the LLM if (fileContents.length > 0) { fullInput += "\n\nFile contents:\n"; fileContents.forEach(file => { fullInput += `\n--- ${file.name} ---\n${file.content}\n`; }); } // Show file icons in chat if there are files let displayContent = inputValue; if (currentFileNames.length > 0) { displayContent += "\n\n"; currentFileNames.forEach(fileName => { displayContent += ` Attached file: ${fileName}\n`; }); } // Add the message to the chat UI with just the icons Alpine.store("chat").add("user", displayContent, images, audios); // Update the last message in the store with the full content const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (activeChat && activeChat.history.length > 0) { activeChat.history[activeChat.history.length - 1].content = fullInput; activeChat.updatedAt = Date.now(); } const input = document.getElementById("input"); if (input) input.value = ""; const systemPrompt = activeChat?.systemPrompt || ""; Alpine.nextTick(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }); // Reset token tracking before starting new request requestStartTime = Date.now(); tokensReceived = 0; promptGPT(systemPrompt, fullInput); // Reset file contents and names after sending fileContents = []; currentFileNames = []; images = []; audios = []; imageFileMap.clear(); audioFileMap.clear(); // Clear Alpine.js attachedFiles array const inputContainer = document.querySelector('[x-data*="attachedFiles"]'); if (inputContainer && window.Alpine) { const alpineData = Alpine.$data(inputContainer); if (alpineData && alpineData.attachedFiles) { alpineData.attachedFiles = []; } } // Clear file inputs document.getElementById("input_image").value = null; document.getElementById("input_audio").value = null; document.getElementById("input_file").value = null; } function readInputImage() { if (!this.files || !this.files.length) return; Array.from(this.files).forEach(file => { readInputImageFile(file); }); } function readInputImageFile(file) { const FR = new FileReader(); FR.addEventListener("load", function(evt) { const dataURL = evt.target.result; images.push(dataURL); imageFileMap.set(file.name, dataURL); }); FR.readAsDataURL(file); } function readInputAudio() { if (!this.files || !this.files.length) return; Array.from(this.files).forEach(file => { readInputAudioFile(file); }); } function readInputAudioFile(file) { const FR = new FileReader(); FR.addEventListener("load", function(evt) { const dataURL = evt.target.result; audios.push(dataURL); audioFileMap.set(file.name, dataURL); }); FR.readAsDataURL(file); } async function promptGPT(systemPrompt, input) { const chatStore = Alpine.store("chat"); const activeChat = chatStore.activeChat(); if (!activeChat) { console.error('No active chat'); return; } const model = activeChat.model || document.getElementById("chat-model").value; const mcpMode = activeChat.mcpMode || false; // Reset current request usage tracking for new request if (activeChat.tokenUsage) { activeChat.tokenUsage.currentRequest = null; } // Store the chat ID for this request so we can track it even if user switches chats const chatId = activeChat.id; toggleLoader(true, chatId); messages = chatStore.messages(); // if systemPrompt isn't empty, push it at the start of messages if (systemPrompt) { messages.unshift({ role: "system", content: systemPrompt }); } // loop all messages, and check if there are images or audios. If there are, we need to change the content field messages.forEach((message) => { if ((message.image && message.image.length > 0) || (message.audio && message.audio.length > 0)) { // The content field now becomes an array message.content = [ { "type": "text", "text": message.content } ] if (message.image && message.image.length > 0) { message.image.forEach(img => { message.content.push( { "type": "image_url", "image_url": { "url": img, } } ); }); delete message.image; } if (message.audio && message.audio.length > 0) { message.audio.forEach(aud => { message.content.push( { "type": "audio_url", "audio_url": { "url": aud, } } ); }); delete message.audio; } } }); // reset the form and the files (already done in processAndSendMessage) // images, audios, and file inputs are cleared after sending // Choose endpoint based on MCP mode const endpoint = mcpMode ? "v1/mcp/chat/completions" : "v1/chat/completions"; const requestBody = { model: model, messages: messages, }; // Add stream parameter for both regular chat and MCP (MCP now supports SSE streaming) requestBody.stream = true; // Add generation parameters if they are set (null means use default) if (activeChat.temperature !== null && activeChat.temperature !== undefined) { requestBody.temperature = activeChat.temperature; } if (activeChat.topP !== null && activeChat.topP !== undefined) { requestBody.top_p = activeChat.topP; } if (activeChat.topK !== null && activeChat.topK !== undefined) { requestBody.top_k = activeChat.topK; } let response; try { // Create AbortController for timeout handling and stop button const controller = new AbortController(); // Store per-chat so switching chats doesn't abort this request const requestStartTime = Date.now(); activeRequests.set(chatId, { controller: controller, reader: null, startTime: requestStartTime, tokensReceived: 0, interval: null, maxTokensPerSecond: 0 }); // Update reactive tracking for UI indicators updateRequestTracking(chatId, true); // Also store globally for stop button (only for active chat) currentAbortController = controller; // Start tokens/second interval now that the request is created // Try to start immediately, and also schedule a retry in case Alpine isn't ready startTokensPerSecondInterval(); setTimeout(() => { // Retry in case the first attempt failed due to timing if (!tokensPerSecondInterval) { startTokensPerSecondInterval(); } }, 200); const timeoutId = setTimeout(() => controller.abort(), mcpMode ? 300000 : 30000); // 5 minutes for MCP, 30 seconds for regular response = await fetch(endpoint, { method: "POST", headers: { "Content-Type": "application/json", "Accept": "application/json", }, body: JSON.stringify(requestBody), signal: controller.signal }); clearTimeout(timeoutId); } catch (error) { // Don't show error if request was aborted by user (stop button) if (error.name === 'AbortError') { // Check if this was a user-initiated abort (stop button was clicked) // If currentAbortController is null, it means stopRequest() was called and already handled the UI if (!currentAbortController) { // User clicked stop button - error message already shown by stopRequest() return; } else { // Timeout error (controller was aborted by timeout, not user) chatStore.add( "assistant", `Request timeout: MCP processing is taking longer than expected. Please try again.`, null, null, chatId ); } } else { chatStore.add( "assistant", `Network Error: ${error.message}`, null, null, chatId ); } toggleLoader(false, chatId); activeRequests.delete(chatId); updateRequestTracking(chatId, false); const activeChat = chatStore.activeChat(); if (activeChat && activeChat.id === chatId) { currentAbortController = null; } return; } if (!response.ok) { chatStore.add( "assistant", `Error: POST ${endpoint} ${response.status}`, null, null, chatId ); toggleLoader(false, chatId); activeRequests.delete(chatId); updateRequestTracking(chatId, false); const activeChat = chatStore.activeChat(); if (activeChat && activeChat.id === chatId) { currentAbortController = null; } return; } // Handle streaming response (both regular and MCP mode now use SSE) if (mcpMode) { // Handle MCP SSE streaming with new event types const reader = response.body ?.pipeThrough(new TextDecoderStream()) .getReader(); if (!reader) { chatStore.add( "assistant", `Error: Failed to decode MCP API response`, null, null, chatId ); toggleLoader(false, chatId); activeRequests.delete(chatId); return; } // Store reader per-chat and globally const mcpRequest = activeRequests.get(chatId); if (mcpRequest) { mcpRequest.reader = reader; // Ensure tracking is updated when reader is set updateRequestTracking(chatId, true); } currentReader = reader; let buffer = ""; let assistantContent = ""; let assistantContentBuffer = []; let thinkingContent = ""; let isThinking = false; let lastAssistantMessageIndex = -1; let lastThinkingMessageIndex = -1; let lastThinkingScrollTime = 0; const THINKING_SCROLL_THROTTLE = 200; // Throttle scrolling to every 200ms try { while (true) { const { value, done } = await reader.read(); if (done) break; // Check if chat still exists and is still the target chat (user might have switched) const currentChat = chatStore.getChat(chatId); if (!currentChat) { // Chat was deleted, abort break; } const targetHistory = currentChat.history; buffer += value; let lines = buffer.split("\n"); buffer = lines.pop(); // Retain any incomplete line in the buffer lines.forEach((line) => { if (line.length === 0 || line.startsWith(":")) return; if (line === "data: [DONE]") { return; } if (line.startsWith("data: ")) { try { const eventData = JSON.parse(line.substring(6)); // Handle different event types switch (eventData.type) { case "reasoning": if (eventData.content) { // Insert reasoning before assistant message if it exists if (lastAssistantMessageIndex >= 0 && targetHistory[lastAssistantMessageIndex]?.role === "assistant") { targetHistory.splice(lastAssistantMessageIndex, 0, { role: "reasoning", content: eventData.content, html: DOMPurify.sanitize(marked.parse(eventData.content)), image: [], audio: [], expanded: false // Reasoning is always collapsed }); lastAssistantMessageIndex++; // Adjust index since we inserted // Scroll smoothly after adding reasoning setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 100); } else { // No assistant message yet, just add normally chatStore.add("reasoning", eventData.content, null, null, chatId); } } break; case "tool_call": if (eventData.name) { // Store as JSON for better formatting const toolCallData = { name: eventData.name, arguments: eventData.arguments || {}, reasoning: eventData.reasoning || "" }; chatStore.add("tool_call", JSON.stringify(toolCallData, null, 2), null, null, chatId); // Scroll smoothly after adding tool call setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 100); } break; case "tool_result": if (eventData.name) { // Store as JSON for better formatting const toolResultData = { name: eventData.name, result: eventData.result || "" }; chatStore.add("tool_result", JSON.stringify(toolResultData, null, 2), null, null, chatId); // Scroll smoothly after adding tool result setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 100); } break; case "status": // Status messages can be logged but not necessarily displayed console.log("[MCP Status]", eventData.message); break; case "assistant": if (eventData.content) { assistantContent += eventData.content; const contentChunk = eventData.content; // Count tokens for rate calculation (per chat) const request = activeRequests.get(chatId); if (request) { request.tokensReceived += Math.ceil(contentChunk.length / 4); } // Only update display if this is the active chat (interval will handle it) // Don't call updateTokensPerSecond here to avoid unnecessary updates // Check for thinking tags in the chunk (incremental detection) if (contentChunk.includes("") || contentChunk.includes("")) { isThinking = true; thinkingContent = ""; lastThinkingMessageIndex = -1; } if (contentChunk.includes("") || contentChunk.includes("")) { isThinking = false; // When closing tag is detected, process the accumulated thinking content if (thinkingContent.trim()) { // Extract just the thinking part from the accumulated content const thinkingMatch = thinkingContent.match(/<(?:thinking|redacted_reasoning)>(.*?)<\/(?:thinking|redacted_reasoning)>/s); if (thinkingMatch && thinkingMatch[1]) { const extractedThinking = thinkingMatch[1]; const currentChat = chatStore.getChat(chatId); if (!currentChat) break; // Chat was deleted const isMCPMode = currentChat.mcpMode || false; const shouldExpand = !isMCPMode; // Expanded in non-MCP mode, collapsed in MCP mode if (lastThinkingMessageIndex === -1) { // Insert thinking before the last assistant message if it exists if (lastAssistantMessageIndex >= 0 && targetHistory[lastAssistantMessageIndex]?.role === "assistant") { // Insert before assistant message targetHistory.splice(lastAssistantMessageIndex, 0, { role: "thinking", content: extractedThinking, html: DOMPurify.sanitize(marked.parse(extractedThinking)), image: [], audio: [], expanded: shouldExpand }); lastThinkingMessageIndex = lastAssistantMessageIndex; lastAssistantMessageIndex++; // Adjust index since we inserted } else { // No assistant message yet, just add normally chatStore.add("thinking", extractedThinking, null, null, chatId); lastThinkingMessageIndex = targetHistory.length - 1; } } else { // Update existing thinking message const lastMessage = targetHistory[lastThinkingMessageIndex]; if (lastMessage && lastMessage.role === "thinking") { lastMessage.content = extractedThinking; lastMessage.html = DOMPurify.sanitize(marked.parse(extractedThinking)); } } // Scroll when thinking is finalized in non-MCP mode if (!isMCPMode) { setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 50); } } thinkingContent = ""; } } // Handle content based on thinking state if (isThinking) { thinkingContent += contentChunk; const currentChat = chatStore.getChat(chatId); if (!currentChat) break; // Chat was deleted const isMCPMode = currentChat.mcpMode || false; const shouldExpand = !isMCPMode; // Expanded in non-MCP mode, collapsed in MCP mode // Update the last thinking message or create a new one (incremental) if (lastThinkingMessageIndex === -1) { // Insert thinking before the last assistant message if it exists if (lastAssistantMessageIndex >= 0 && targetHistory[lastAssistantMessageIndex]?.role === "assistant") { // Insert before assistant message targetHistory.splice(lastAssistantMessageIndex, 0, { role: "thinking", content: thinkingContent, html: DOMPurify.sanitize(marked.parse(thinkingContent)), image: [], audio: [], expanded: shouldExpand }); lastThinkingMessageIndex = lastAssistantMessageIndex; lastAssistantMessageIndex++; // Adjust index since we inserted } else { // No assistant message yet, just add normally chatStore.add("thinking", thinkingContent, null, null, chatId); lastThinkingMessageIndex = targetHistory.length - 1; } } else { // Update existing thinking message const lastMessage = targetHistory[lastThinkingMessageIndex]; if (lastMessage && lastMessage.role === "thinking") { lastMessage.content = thinkingContent; lastMessage.html = DOMPurify.sanitize(marked.parse(thinkingContent)); } } // Scroll when thinking is updated in non-MCP mode (throttled) if (!isMCPMode) { const now = Date.now(); if (now - lastThinkingScrollTime > THINKING_SCROLL_THROTTLE) { lastThinkingScrollTime = now; setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 100); } } } else { // Regular assistant content - buffer it for batch processing assistantContentBuffer.push(contentChunk); } } break; case "error": chatStore.add( "assistant", `MCP Error: ${eventData.message}`, null, null, chatId ); break; } } catch (error) { console.error("Failed to parse MCP event:", line, error); } } }); // Efficiently update assistant message in batch if (assistantContentBuffer.length > 0) { const regularContent = assistantContentBuffer.join(""); // Process any thinking tags that might be in the accumulated content // This handles cases where tags are split across chunks const { regularContent: processedRegular, thinkingContent: processedThinking } = processThinkingTags(regularContent); // Update or create assistant message with processed regular content const currentChat = chatStore.getChat(chatId); if (!currentChat) break; // Chat was deleted if (lastAssistantMessageIndex === -1) { if (processedRegular && processedRegular.trim()) { chatStore.add("assistant", processedRegular, null, null, chatId); lastAssistantMessageIndex = targetHistory.length - 1; } } else { const lastMessage = targetHistory[lastAssistantMessageIndex]; if (lastMessage && lastMessage.role === "assistant") { lastMessage.content = (lastMessage.content || "") + (processedRegular || ""); lastMessage.html = DOMPurify.sanitize(marked.parse(lastMessage.content)); } } // Add any extracted thinking content from the processed buffer BEFORE assistant message if (processedThinking && processedThinking.trim()) { const isMCPMode = currentChat.mcpMode || false; const shouldExpand = !isMCPMode; // Expanded in non-MCP mode, collapsed in MCP mode // Insert thinking before assistant message if it exists if (lastAssistantMessageIndex >= 0 && targetHistory[lastAssistantMessageIndex]?.role === "assistant") { targetHistory.splice(lastAssistantMessageIndex, 0, { role: "thinking", content: processedThinking, html: DOMPurify.sanitize(marked.parse(processedThinking)), image: [], audio: [], expanded: shouldExpand }); lastAssistantMessageIndex++; // Adjust index since we inserted } else { // No assistant message yet, just add normally chatStore.add("thinking", processedThinking, null, null, chatId); } } assistantContentBuffer = []; } } // Final assistant content flush if any data remains if (assistantContentBuffer.length > 0) { const regularContent = assistantContentBuffer.join(""); // Process any remaining thinking tags that might be in the buffer const { regularContent: processedRegular, thinkingContent: processedThinking } = processThinkingTags(regularContent); const currentChat = chatStore.getChat(chatId); if (!currentChat) { // Chat was deleted, cleanup and exit activeRequests.delete(chatId); updateRequestTracking(chatId, false); return; } const targetHistory = currentChat.history; // First, add any extracted thinking content BEFORE assistant message if (processedThinking && processedThinking.trim()) { const isMCPMode = currentChat.mcpMode || false; const shouldExpand = !isMCPMode; // Expanded in non-MCP mode, collapsed in MCP mode // Insert thinking before assistant message if it exists if (lastAssistantMessageIndex >= 0 && targetHistory[lastAssistantMessageIndex]?.role === "assistant") { targetHistory.splice(lastAssistantMessageIndex, 0, { role: "thinking", content: processedThinking, html: DOMPurify.sanitize(marked.parse(processedThinking)), image: [], audio: [], expanded: shouldExpand }); lastAssistantMessageIndex++; // Adjust index since we inserted } else { // No assistant message yet, just add normally chatStore.add("thinking", processedThinking, null, null, chatId); } } // Then update or create assistant message if (lastAssistantMessageIndex !== -1) { const lastMessage = targetHistory[lastAssistantMessageIndex]; if (lastMessage && lastMessage.role === "assistant") { lastMessage.content = (lastMessage.content || "") + (processedRegular || ""); lastMessage.html = DOMPurify.sanitize(marked.parse(lastMessage.content)); } } else if (processedRegular && processedRegular.trim()) { chatStore.add("assistant", processedRegular, null, null, chatId); lastAssistantMessageIndex = targetHistory.length - 1; } } // Final thinking content flush if any data remains (from incremental detection) const finalChat = chatStore.getChat(chatId); if (finalChat && thinkingContent.trim() && lastThinkingMessageIndex === -1) { const finalHistory = finalChat.history; // Extract thinking content if tags are present const thinkingMatch = thinkingContent.match(/<(?:thinking|redacted_reasoning)>(.*?)<\/(?:thinking|redacted_reasoning)>/s); if (thinkingMatch && thinkingMatch[1]) { const isMCPMode = finalChat.mcpMode || false; const shouldExpand = !isMCPMode; // Expanded in non-MCP mode, collapsed in MCP mode // Insert thinking before assistant message if it exists if (lastAssistantMessageIndex >= 0 && finalHistory[lastAssistantMessageIndex]?.role === "assistant") { finalHistory.splice(lastAssistantMessageIndex, 0, { role: "thinking", content: thinkingMatch[1], html: DOMPurify.sanitize(marked.parse(thinkingMatch[1])), image: [], audio: [], expanded: shouldExpand }); } else { // No assistant message yet, just add normally chatStore.add("thinking", thinkingMatch[1], null, null, chatId); } } else { chatStore.add("thinking", thinkingContent, null, null, chatId); } } // Final pass: process the entire assistantContent to catch any missed thinking tags // This ensures we don't miss tags that were split across chunks if (finalChat && assistantContent.trim()) { const finalHistory = finalChat.history; const { regularContent: finalRegular, thinkingContent: finalThinking } = processThinkingTags(assistantContent); // Update assistant message with final processed content (without thinking tags) if (finalRegular && finalRegular.trim()) { if (lastAssistantMessageIndex !== -1) { const lastMessage = finalHistory[lastAssistantMessageIndex]; if (lastMessage && lastMessage.role === "assistant") { lastMessage.content = finalRegular; lastMessage.html = DOMPurify.sanitize(marked.parse(lastMessage.content)); } } else { chatStore.add("assistant", finalRegular, null, null, chatId); } } // Add any extracted thinking content (only if not already added) if (finalThinking && finalThinking.trim()) { const hasThinking = finalHistory.some(msg => msg.role === "thinking" && msg.content.trim() === finalThinking.trim() ); if (!hasThinking) { chatStore.add("thinking", finalThinking, null, null, chatId); } } } // Cleanup request tracking activeRequests.delete(chatId); updateRequestTracking(chatId, false); // Highlight all code blocks once at the end hljs.highlightAll(); } catch (error) { // Don't show error if request was aborted by user if (error.name !== 'AbortError' || !currentAbortController) { const errorChat = chatStore.getChat(chatId); if (errorChat) { chatStore.add( "assistant", `Error: Failed to process MCP stream`, null, null, chatId ); } } } finally { // Perform any cleanup if necessary if (reader) { reader.releaseLock(); } // Only clear global references if this was the active chat's request const activeChat = chatStore.activeChat(); if (activeChat && activeChat.id === chatId) { currentReader = null; currentAbortController = null; toggleLoader(false, chatId); } // Cleanup per-chat tracking activeRequests.delete(chatId); updateRequestTracking(chatId, false); } } else { // Handle regular streaming response const reader = response.body ?.pipeThrough(new TextDecoderStream()) .getReader(); if (!reader) { chatStore.add( "assistant", `Error: Failed to decode API response`, null, null, chatId ); toggleLoader(false, chatId); activeRequests.delete(chatId); return; } // Store reader per-chat and globally const request = activeRequests.get(chatId); if (request) { request.reader = reader; // Ensure tracking is updated when reader is set updateRequestTracking(chatId, true); // Ensure interval is running (in case it wasn't started earlier) startTokensPerSecondInterval(); } currentReader = reader; // Get target chat for this request let targetChat = chatStore.getChat(chatId); if (!targetChat) { // Chat was deleted activeRequests.delete(chatId); updateRequestTracking(chatId, false); return; } // Function to add content to the chat and handle DOM updates efficiently const addToChat = (token) => { const currentChat = chatStore.getChat(chatId); if (!currentChat) return; // Chat was deleted chatStore.add("assistant", token, null, null, chatId); // Count tokens for rate calculation (per chat) const request = activeRequests.get(chatId); if (request) { const tokenCount = Math.ceil(token.length / 4); request.tokensReceived += tokenCount; } // Only update display if this is the active chat (interval will handle it) // Don't call updateTokensPerSecond here to avoid unnecessary updates }; let buffer = ""; let contentBuffer = []; let thinkingContent = ""; let isThinking = false; let lastThinkingMessageIndex = -1; let lastThinkingScrollTime = 0; const THINKING_SCROLL_THROTTLE = 200; // Throttle scrolling to every 200ms try { while (true) { const { value, done } = await reader.read(); if (done) break; // Check if chat still exists targetChat = chatStore.getChat(chatId); if (!targetChat) { // Chat was deleted, abort break; } const targetHistory = targetChat.history; buffer += value; let lines = buffer.split("\n"); buffer = lines.pop(); // Retain any incomplete line in the buffer lines.forEach((line) => { if (line.length === 0 || line.startsWith(":")) return; if (line === "data: [DONE]") { return; } if (line.startsWith("data: ")) { try { const jsonData = JSON.parse(line.substring(6)); // Update token usage if present (for the chat that initiated this request) if (jsonData.usage) { chatStore.updateTokenUsage(jsonData.usage, chatId); } const token = jsonData.choices[0].delta.content; if (token) { // Check for thinking tags if (token.includes("") || token.includes("")) { isThinking = true; thinkingContent = ""; lastThinkingMessageIndex = -1; return; } if (token.includes("") || token.includes("")) { isThinking = false; if (thinkingContent.trim()) { // Only add the final thinking message if we don't already have one if (lastThinkingMessageIndex === -1) { chatStore.add("thinking", thinkingContent, null, null, chatId); } } return; } // Handle content based on thinking state if (isThinking) { thinkingContent += token; // Count tokens for rate calculation (per chat) const request = activeRequests.get(chatId); if (request) { request.tokensReceived += Math.ceil(token.length / 4); } // Only update display if this is the active chat (interval will handle it) // Don't call updateTokensPerSecond here to avoid unnecessary updates // Update the last thinking message or create a new one if (lastThinkingMessageIndex === -1) { // Create new thinking message chatStore.add("thinking", thinkingContent, null, null, chatId); const targetChat = chatStore.getChat(chatId); lastThinkingMessageIndex = targetChat ? targetChat.history.length - 1 : -1; } else { // Update existing thinking message const currentChat = chatStore.getChat(chatId); if (currentChat && lastThinkingMessageIndex >= 0) { const lastMessage = currentChat.history[lastThinkingMessageIndex]; if (lastMessage && lastMessage.role === "thinking") { lastMessage.content = thinkingContent; lastMessage.html = DOMPurify.sanitize(marked.parse(thinkingContent)); } } } // Scroll when thinking is updated (throttled) const now = Date.now(); if (now - lastThinkingScrollTime > THINKING_SCROLL_THROTTLE) { lastThinkingScrollTime = now; setTimeout(() => { // Scroll main chat container const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } // Scroll thinking box to bottom if it's expanded and scrollable scrollThinkingBoxToBottom(); }, 100); } } else { contentBuffer.push(token); } } } catch (error) { console.error("Failed to parse line:", line, error); } } }); // Efficiently update the chat in batch if (contentBuffer.length > 0) { addToChat(contentBuffer.join("")); contentBuffer = []; // Scroll when assistant content is updated (this will also show thinking messages above) setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 50); } } // Final content flush if any data remains if (contentBuffer.length > 0) { addToChat(contentBuffer.join("")); } const finalChat = chatStore.getChat(chatId); if (finalChat && thinkingContent.trim() && lastThinkingMessageIndex === -1) { chatStore.add("thinking", thinkingContent, null, null, chatId); } // Highlight all code blocks once at the end hljs.highlightAll(); } catch (error) { // Don't show error if request was aborted by user if (error.name !== 'AbortError' || !currentAbortController) { const currentChat = chatStore.getChat(chatId); if (currentChat) { chatStore.add( "assistant", `Error: Failed to process stream`, null, null, chatId ); } } } finally { // Perform any cleanup if necessary if (reader) { reader.releaseLock(); } // Only clear global references if this was the active chat's request const activeChat = chatStore.activeChat(); if (activeChat && activeChat.id === chatId) { currentReader = null; currentAbortController = null; toggleLoader(false, chatId); } // Cleanup per-chat tracking activeRequests.delete(chatId); updateRequestTracking(chatId, false); } } // Remove class "loader" from the element with "loader" id // Only toggle loader off if this was the active chat const finalActiveChat = chatStore.activeChat(); if (finalActiveChat && finalActiveChat.id === chatId) { toggleLoader(false, chatId); } // scroll to the bottom of the chat consistently setTimeout(() => { const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } }, 100); // set focus to the input document.getElementById("input").focus(); } document.getElementById("system_prompt").addEventListener("submit", submitSystemPrompt); document.getElementById("prompt").addEventListener("submit", submitPrompt); document.getElementById("input").focus(); storesystemPrompt = localStorage.getItem("system_prompt"); if (storesystemPrompt) { document.getElementById("systemPrompt").value = storesystemPrompt; } else { document.getElementById("systemPrompt").value = null; } marked.setOptions({ highlight: function (code) { return hljs.highlightAuto(code).value; }, }); // Alpine store is now initialized in chat.html inline script to ensure it's available before Alpine processes the DOM // Only initialize if not already initialized (to avoid duplicate initialization) document.addEventListener("alpine:init", () => { // Check if store already exists (initialized in chat.html) if (!Alpine.store("chat")) { // Fallback initialization (should not be needed if chat.html loads correctly) // This matches the structure in chat.html function generateChatId() { return "chat_" + Date.now() + "_" + Math.random().toString(36).substr(2, 9); } function getCurrentModel() { const modelInput = document.getElementById("chat-model"); return modelInput ? modelInput.value : ""; } Alpine.store("chat", { chats: [], activeChatId: null, chatIdCounter: 0, languages: [undefined], activeRequestIds: [], // Track chat IDs with active requests for UI reactivity activeChat() { if (!this.activeChatId) return null; return this.chats.find(c => c.id === this.activeChatId) || null; }, getChat(chatId) { return this.chats.find(c => c.id === chatId) || null; }, createChat(model, systemPrompt, mcpMode) { const chatId = generateChatId(); const now = Date.now(); const chat = { id: chatId, name: "New Chat", model: model || getCurrentModel() || "", history: [], systemPrompt: systemPrompt || "", mcpMode: mcpMode || false, tokenUsage: { promptTokens: 0, completionTokens: 0, totalTokens: 0, currentRequest: null }, contextSize: null, createdAt: now, updatedAt: now }; this.chats.push(chat); this.activeChatId = chatId; return chat; }, switchChat(chatId) { if (this.chats.find(c => c.id === chatId)) { this.activeChatId = chatId; return true; } return false; }, deleteChat(chatId) { const index = this.chats.findIndex(c => c.id === chatId); if (index === -1) return false; this.chats.splice(index, 1); if (this.activeChatId === chatId) { if (this.chats.length > 0) { this.activeChatId = this.chats[0].id; } else { this.createChat(); } } return true; }, updateChatName(chatId, name) { const chat = this.getChat(chatId); if (chat) { chat.name = name || "New Chat"; chat.updatedAt = Date.now(); return true; } return false; }, clear() { const chat = this.activeChat(); if (chat) { chat.history.length = 0; chat.tokenUsage = { promptTokens: 0, completionTokens: 0, totalTokens: 0, currentRequest: null }; chat.updatedAt = Date.now(); } }, updateTokenUsage(usage, targetChatId = null) { // If targetChatId is provided, update that chat, otherwise use active chat // This ensures token usage updates go to the chat that initiated the request const chat = targetChatId ? this.getChat(targetChatId) : this.activeChat(); if (!chat) return; if (usage) { const currentRequest = chat.tokenUsage.currentRequest || { promptTokens: 0, completionTokens: 0, totalTokens: 0 }; const isNewUsage = (usage.prompt_tokens !== undefined && usage.prompt_tokens > currentRequest.promptTokens) || (usage.completion_tokens !== undefined && usage.completion_tokens > currentRequest.completionTokens) || (usage.total_tokens !== undefined && usage.total_tokens > currentRequest.totalTokens); if (isNewUsage) { chat.tokenUsage.promptTokens = chat.tokenUsage.promptTokens - currentRequest.promptTokens + (usage.prompt_tokens || 0); chat.tokenUsage.completionTokens = chat.tokenUsage.completionTokens - currentRequest.completionTokens + (usage.completion_tokens || 0); chat.tokenUsage.totalTokens = chat.tokenUsage.totalTokens - currentRequest.totalTokens + (usage.total_tokens || 0); chat.tokenUsage.currentRequest = { promptTokens: usage.prompt_tokens || 0, completionTokens: usage.completion_tokens || 0, totalTokens: usage.total_tokens || 0 }; chat.updatedAt = Date.now(); } } }, getRemainingTokens() { const chat = this.activeChat(); if (!chat || !chat.contextSize) return null; return Math.max(0, chat.contextSize - chat.tokenUsage.totalTokens); }, getContextUsagePercent() { const chat = this.activeChat(); if (!chat || !chat.contextSize) return null; return Math.min(100, (chat.tokenUsage.totalTokens / chat.contextSize) * 100); }, // Check if a chat has an active request (for UI indicators) hasActiveRequest(chatId) { if (!chatId) return false; // Use reactive array for Alpine.js reactivity return this.activeRequestIds.includes(chatId); }, // Update active request tracking (called from chat.js) updateActiveRequestTracking(chatId, isActive) { if (isActive) { if (!this.activeRequestIds.includes(chatId)) { this.activeRequestIds.push(chatId); } } else { const index = this.activeRequestIds.indexOf(chatId); if (index > -1) { this.activeRequestIds.splice(index, 1); } } }, add(role, content, image, audio, targetChatId = null) { // If targetChatId is provided, add to that chat, otherwise use active chat const chat = targetChatId ? this.getChat(targetChatId) : this.activeChat(); if (!chat) return; const N = chat.history.length - 1; if (role === "thinking" || role === "reasoning") { let c = ""; const lines = content.split("\n"); lines.forEach((line) => { c += DOMPurify.sanitize(marked.parse(line)); }); chat.history.push({ role, content, html: c, image, audio }); } else if (chat.history.length && chat.history[N].role === role) { chat.history[N].content += content; chat.history[N].html = DOMPurify.sanitize( marked.parse(chat.history[N].content) ); if (image && image.length > 0) { chat.history[N].image = [...(chat.history[N].image || []), ...image]; } if (audio && audio.length > 0) { chat.history[N].audio = [...(chat.history[N].audio || []), ...audio]; } } else { let c = ""; const lines = content.split("\n"); lines.forEach((line) => { c += DOMPurify.sanitize(marked.parse(line)); }); chat.history.push({ role, content, html: c, image: image || [], audio: audio || [] }); if (role === "user" && chat.name === "New Chat" && content.trim()) { const name = content.trim().substring(0, 50); chat.name = name.length < content.trim().length ? name + "..." : name; } } chat.updatedAt = Date.now(); const chatContainer = document.getElementById('chat'); if (chatContainer) { chatContainer.scrollTo({ top: chatContainer.scrollHeight, behavior: 'smooth' }); } if (role === "thinking" || role === "reasoning") { setTimeout(() => { if (typeof window.scrollThinkingBoxToBottom === 'function') { window.scrollThinkingBoxToBottom(); } }, 100); } const parser = new DOMParser(); const html = parser.parseFromString( chat.history[chat.history.length - 1].html, "text/html" ); const code = html.querySelectorAll("pre code"); if (!code.length) return; code.forEach((el) => { const language = el.className.split("language-")[1]; if (this.languages.includes(language)) return; const script = document.createElement("script"); script.src = `https://cdn.jsdelivr.net/gh/highlightjs/cdn-release@11.8.0/build/languages/${language}.min.js`; document.head.appendChild(script); this.languages.push(language); }); }, messages() { const chat = this.activeChat(); if (!chat) return []; return chat.history.map((message) => ({ role: message.role, content: message.content, image: message.image, audio: message.audio, })); }, // Getter for active chat history to ensure reactivity get activeHistory() { const chat = this.activeChat(); return chat ? chat.history : []; }, }); } }); // Check for message from index page on load and initialize chats document.addEventListener('DOMContentLoaded', function() { // Wait for Alpine to be ready setTimeout(() => { if (!window.Alpine || !Alpine.store("chat")) { console.error('Alpine store not initialized'); return; } const chatStore = Alpine.store("chat"); // Check for message from index page FIRST - if present, create new chat const chatData = localStorage.getItem('localai_index_chat_data'); let shouldCreateNewChat = false; let indexChatData = null; if (chatData) { try { indexChatData = JSON.parse(chatData); shouldCreateNewChat = true; // We have data from index, create new chat } catch (error) { console.error('Error parsing chat data from index:', error); localStorage.removeItem('localai_index_chat_data'); } } // Load chats from storage FIRST (but don't set active yet if we're creating new from index) const storedData = loadChatsFromStorage(); if (storedData && storedData.chats && storedData.chats.length > 0) { // Restore chats from storage - clear existing and push new ones to maintain reactivity chatStore.chats.length = 0; storedData.chats.forEach(chat => { chatStore.chats.push(chat); }); // Don't set activeChatId yet if we're creating a new chat from index if (!shouldCreateNewChat) { chatStore.activeChatId = storedData.activeChatId || storedData.chats[0].id; // Ensure active chat exists if (!chatStore.activeChat()) { chatStore.activeChatId = storedData.chats[0].id; } } } if (shouldCreateNewChat) { // Create a new chat with the model from URL (which matches the selected model from index) const currentModel = document.getElementById("chat-model")?.value || ""; // Check URL parameter for MCP mode (takes precedence over localStorage) const urlParams = new URLSearchParams(window.location.search); const mcpFromUrl = urlParams.get('mcp') === 'true'; const newChat = chatStore.createChat(currentModel, "", mcpFromUrl || indexChatData.mcpMode || false); // Update context size from template if available const contextSizeInput = document.getElementById("chat-model"); if (contextSizeInput && contextSizeInput.dataset.contextSize) { const contextSize = parseInt(contextSizeInput.dataset.contextSize); newChat.contextSize = contextSize; } // Set the message and files const input = document.getElementById('input'); if (input && indexChatData.message) { input.value = indexChatData.message; // Process files if any if (indexChatData.imageFiles && indexChatData.imageFiles.length > 0) { indexChatData.imageFiles.forEach(file => { images.push(file.data); }); } if (indexChatData.audioFiles && indexChatData.audioFiles.length > 0) { indexChatData.audioFiles.forEach(file => { audios.push(file.data); }); } if (indexChatData.textFiles && indexChatData.textFiles.length > 0) { indexChatData.textFiles.forEach(file => { fileContents.push({ name: file.name, content: file.data }); currentFileNames.push(file.name); }); } // Clear localStorage localStorage.removeItem('localai_index_chat_data'); // Save the new chat saveChatsToStorage(); // Update UI to reflect new active chat updateUIForActiveChat(); // Auto-submit after a short delay to ensure everything is ready setTimeout(() => { if (input.value.trim()) { processAndSendMessage(input.value); } }, 500); } else { // No message, but might have mcpMode from URL - clear localStorage localStorage.removeItem('localai_index_chat_data'); // If MCP mode was set from URL, ensure it's enabled const urlParams = new URLSearchParams(window.location.search); if (urlParams.get('mcp') === 'true' && newChat) { newChat.mcpMode = true; saveChatsToStorage(); updateUIForActiveChat(); } saveChatsToStorage(); updateUIForActiveChat(); } } else { // Normal flow: create default chat if none exist if (!storedData || !storedData.chats || storedData.chats.length === 0) { const currentModel = document.getElementById("chat-model")?.value || ""; const oldSystemPrompt = localStorage.getItem(SYSTEM_PROMPT_STORAGE_KEY); // Check URL parameter for MCP mode const urlParams = new URLSearchParams(window.location.search); const mcpFromUrl = urlParams.get('mcp') === 'true'; chatStore.createChat(currentModel, oldSystemPrompt || "", mcpFromUrl); // Remove old system prompt key after migration if (oldSystemPrompt) { localStorage.removeItem(SYSTEM_PROMPT_STORAGE_KEY); } } else { // Existing chats loaded - check if we need to create a new chat for the model in URL const urlModel = document.getElementById("chat-model")?.value || ""; const activeChat = chatStore.activeChat(); const shouldCreateNewChat = sessionStorage.getItem('localai_create_new_chat') === 'true'; // Clear the flag after reading it if (shouldCreateNewChat) { sessionStorage.removeItem('localai_create_new_chat'); } // If we should create a new chat (from manage.html) or URL model doesn't match active chat, create new chat // This handles navigation from manage.html or direct links to /chat/MODEL_NAME if (urlModel && urlModel.trim() && (shouldCreateNewChat || (activeChat && activeChat.model !== urlModel) || !activeChat)) { // Create a new chat with the model from URL const urlParams = new URLSearchParams(window.location.search); const mcpFromUrl = urlParams.get('mcp') === 'true'; const newChat = chatStore.createChat(urlModel, "", mcpFromUrl); // Update context size from template if available const contextSizeInput = document.getElementById("chat-model"); if (contextSizeInput && contextSizeInput.dataset.contextSize) { const contextSize = parseInt(contextSizeInput.dataset.contextSize); if (!isNaN(contextSize)) { newChat.contextSize = contextSize; } } saveChatsToStorage(); updateUIForActiveChat(); } else { // Check URL parameter for MCP mode (update existing active chat) const urlParams = new URLSearchParams(window.location.search); if (urlParams.get('mcp') === 'true') { if (activeChat) { activeChat.mcpMode = true; saveChatsToStorage(); } } } } // Update context size from template if available (for existing active chat) const contextSizeInput = document.getElementById("chat-model"); if (contextSizeInput && contextSizeInput.dataset.contextSize) { const contextSize = parseInt(contextSizeInput.dataset.contextSize); const activeChat = chatStore.activeChat(); if (activeChat && !activeChat.contextSize) { activeChat.contextSize = contextSize; } } // Update UI to reflect active chat updateUIForActiveChat(); } // Save initial state saveChatsToStorage(); }, 300); });