mirror of
https://github.com/open-webui/open-webui.git
synced 2025-03-27 02:02:31 +01:00
Merge branch 'open-webui:dev' into dev
This commit is contained in:
commit
c9079d4b4d
@ -5,6 +5,12 @@ All notable changes to this project will be documented in this file.
|
||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
|
||||
## [0.5.16] - 2025-02-20
|
||||
|
||||
### Fixed
|
||||
|
||||
- **🔍 Web Search Retrieval Restored**: Resolved a critical issue that broke web search retrieval by reverting deduplication changes, ensuring complete and accurate search results once again.
|
||||
|
||||
## [0.5.15] - 2025-02-20
|
||||
|
||||
### Added
|
||||
|
@ -1714,7 +1714,7 @@ Respond to the user query using the provided context, incorporating inline citat
|
||||
- Respond in the same language as the user's query.
|
||||
- If the context is unreadable or of poor quality, inform the user and provide the best possible answer.
|
||||
- If the answer isn't present in the context but you possess the knowledge, explain this to the user and provide the answer using your own understanding.
|
||||
- **Only include inline citations using [source_id] when a <source_id> tag is explicitly provided in the context.**
|
||||
- **Only include inline citations using [source_id] (e.g., [1], [2]) when a `<source_id>` tag is explicitly provided in the context.**
|
||||
- Do not cite if the <source_id> tag is not provided in the context.
|
||||
- Do not use XML tags in your response.
|
||||
- Ensure citations are concise and directly related to the information provided.
|
||||
@ -1957,7 +1957,7 @@ RAG_WEB_LOADER_ENGINE = PersistentConfig(
|
||||
RAG_WEB_SEARCH_TRUST_ENV = PersistentConfig(
|
||||
"RAG_WEB_SEARCH_TRUST_ENV",
|
||||
"rag.web.search.trust_env",
|
||||
os.getenv("RAG_WEB_SEARCH_TRUST_ENV", False),
|
||||
os.getenv("RAG_WEB_SEARCH_TRUST_ENV", "False").lower() == "true",
|
||||
)
|
||||
|
||||
PLAYWRIGHT_WS_URI = PersistentConfig(
|
||||
|
@ -14,7 +14,8 @@ from langchain_core.documents import Document
|
||||
|
||||
from open_webui.config import VECTOR_DB
|
||||
from open_webui.retrieval.vector.connector import VECTOR_DB_CLIENT
|
||||
from open_webui.utils.misc import get_last_user_message
|
||||
from open_webui.utils.misc import get_last_user_message, calculate_sha256_string
|
||||
|
||||
from open_webui.models.users import UserModel
|
||||
|
||||
from open_webui.env import (
|
||||
@ -178,45 +179,31 @@ def merge_and_sort_query_results(
|
||||
combined_distances = []
|
||||
combined_documents = []
|
||||
combined_metadatas = []
|
||||
combined_ids = []
|
||||
|
||||
for data in query_results:
|
||||
combined_distances.extend(data["distances"][0])
|
||||
combined_documents.extend(data["documents"][0])
|
||||
combined_metadatas.extend(data["metadatas"][0])
|
||||
# DISTINCT(chunk_id,file_id) - in case if id (chunk_ids) become ordinals
|
||||
combined_ids.extend(
|
||||
[
|
||||
f"{id}-{meta['file_id']}"
|
||||
for id, meta in zip(data["ids"][0], data["metadatas"][0])
|
||||
]
|
||||
)
|
||||
|
||||
# Create a list of tuples (distance, document, metadata, ids)
|
||||
combined = list(
|
||||
zip(combined_distances, combined_documents, combined_metadatas, combined_ids)
|
||||
)
|
||||
# Create a list of tuples (distance, document, metadata)
|
||||
combined = list(zip(combined_distances, combined_documents, combined_metadatas))
|
||||
|
||||
# Sort the list based on distances
|
||||
combined.sort(key=lambda x: x[0], reverse=reverse)
|
||||
|
||||
sorted_distances = []
|
||||
sorted_documents = []
|
||||
sorted_metadatas = []
|
||||
# Otherwise we don't have anything :-(
|
||||
if combined:
|
||||
# We don't have anything :-(
|
||||
if not combined:
|
||||
sorted_distances = []
|
||||
sorted_documents = []
|
||||
sorted_metadatas = []
|
||||
else:
|
||||
# Unzip the sorted list
|
||||
all_distances, all_documents, all_metadatas, all_ids = zip(*combined)
|
||||
seen_ids = set()
|
||||
sorted_distances, sorted_documents, sorted_metadatas = zip(*combined)
|
||||
|
||||
# Slicing the lists to include only k elements
|
||||
for index, id in enumerate(all_ids):
|
||||
if id not in seen_ids:
|
||||
sorted_distances.append(all_distances[index])
|
||||
sorted_documents.append(all_documents[index])
|
||||
sorted_metadatas.append(all_metadatas[index])
|
||||
seen_ids.add(id)
|
||||
if len(sorted_distances) >= k:
|
||||
break
|
||||
sorted_distances = list(sorted_distances)[:k]
|
||||
sorted_documents = list(sorted_documents)[:k]
|
||||
sorted_metadatas = list(sorted_metadatas)[:k]
|
||||
|
||||
# Create the output dictionary
|
||||
result = {
|
||||
|
@ -20,6 +20,10 @@ from open_webui.utils.auth import get_admin_user, get_verified_user
|
||||
from open_webui.constants import TASKS
|
||||
|
||||
from open_webui.routers.pipelines import process_pipeline_inlet_filter
|
||||
from open_webui.utils.filter import (
|
||||
get_sorted_filter_ids,
|
||||
process_filter_functions,
|
||||
)
|
||||
from open_webui.utils.task import get_task_model_id
|
||||
|
||||
from open_webui.config import (
|
||||
@ -221,6 +225,12 @@ async def generate_title(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -290,6 +300,12 @@ async def generate_chat_tags(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -356,6 +372,12 @@ async def generate_image_prompt(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -433,6 +455,12 @@ async def generate_queries(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -514,6 +542,12 @@ async def generate_autocompletion(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -584,6 +618,12 @@ async def generate_emoji(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
@ -644,6 +684,12 @@ async def generate_moa_response(
|
||||
},
|
||||
}
|
||||
|
||||
# Process the payload through the pipeline
|
||||
try:
|
||||
payload = await process_pipeline_inlet_filter(request, payload, user, models)
|
||||
except Exception as e:
|
||||
raise e
|
||||
|
||||
try:
|
||||
return await generate_chat_completion(request, form_data=payload, user=user)
|
||||
except Exception as e:
|
||||
|
@ -104,7 +104,7 @@ async def convert_streaming_response_ollama_to_openai(ollama_streaming_response)
|
||||
data = json.loads(data)
|
||||
|
||||
model = data.get("model", "ollama")
|
||||
message_content = data.get("message", {}).get("content", "")
|
||||
message_content = data.get("message", {}).get("content", None)
|
||||
tool_calls = data.get("message", {}).get("tool_calls", None)
|
||||
openai_tool_calls = None
|
||||
|
||||
@ -118,7 +118,7 @@ async def convert_streaming_response_ollama_to_openai(ollama_streaming_response)
|
||||
usage = convert_ollama_usage_to_openai(data)
|
||||
|
||||
data = openai_chat_chunk_message_template(
|
||||
model, message_content if not done else None, openai_tool_calls, usage
|
||||
model, message_content, openai_tool_calls, usage
|
||||
)
|
||||
|
||||
line = f"data: {json.dumps(data)}\n\n"
|
||||
|
4
package-lock.json
generated
4
package-lock.json
generated
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "open-webui",
|
||||
"version": "0.5.15",
|
||||
"version": "0.5.16",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "open-webui",
|
||||
"version": "0.5.15",
|
||||
"version": "0.5.16",
|
||||
"dependencies": {
|
||||
"@codemirror/lang-javascript": "^6.2.2",
|
||||
"@codemirror/lang-python": "^6.1.6",
|
||||
|
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "open-webui",
|
||||
"version": "0.5.15",
|
||||
"version": "0.5.16",
|
||||
"private": true,
|
||||
"scripts": {
|
||||
"dev": "npm run pyodide:fetch && vite dev --host",
|
||||
|
@ -20,7 +20,7 @@
|
||||
"Account Activation Pending": "Activació del compte pendent",
|
||||
"Accurate information": "Informació precisa",
|
||||
"Actions": "Accions",
|
||||
"Activate": "",
|
||||
"Activate": "Activar",
|
||||
"Activate this command by typing \"/{{COMMAND}}\" to chat input.": "Activa aquest comanda escrivint \"{{COMMAND}}\" en el xat",
|
||||
"Active Users": "Usuaris actius",
|
||||
"Add": "Afegir",
|
||||
@ -100,7 +100,7 @@
|
||||
"Audio": "Àudio",
|
||||
"August": "Agost",
|
||||
"Authenticate": "Autenticar",
|
||||
"Authentication": "",
|
||||
"Authentication": "Autenticació",
|
||||
"Auto-Copy Response to Clipboard": "Copiar la resposta automàticament al porta-retalls",
|
||||
"Auto-playback response": "Reproduir la resposta automàticament",
|
||||
"Autocomplete Generation": "Generació automàtica",
|
||||
@ -124,11 +124,11 @@
|
||||
"Beta": "Beta",
|
||||
"Bing Search V7 Endpoint": "Punt de connexió a Bing Search V7",
|
||||
"Bing Search V7 Subscription Key": "Clau de subscripció a Bing Search V7",
|
||||
"Bocha Search API Key": "",
|
||||
"Bocha Search API Key": "Clau API de Bocha Search",
|
||||
"Brave Search API Key": "Clau API de Brave Search",
|
||||
"By {{name}}": "Per {{name}}",
|
||||
"Bypass SSL verification for Websites": "Desactivar la verificació SSL per a l'accés a Internet",
|
||||
"Calendar": "",
|
||||
"Calendar": "Calendari",
|
||||
"Call": "Trucada",
|
||||
"Call feature is not supported when using Web STT engine": "La funció de trucada no s'admet quan s'utilitza el motor Web STT",
|
||||
"Camera": "Càmera",
|
||||
@ -180,13 +180,13 @@
|
||||
"Clone of {{TITLE}}": "Clon de {{TITLE}}",
|
||||
"Close": "Tancar",
|
||||
"Code execution": "Execució de codi",
|
||||
"Code Execution": "",
|
||||
"Code Execution Engine": "",
|
||||
"Code Execution Timeout": "",
|
||||
"Code Execution": "Excució de Codi",
|
||||
"Code Execution Engine": "Motor d'execució de codi",
|
||||
"Code Execution Timeout": "Temps màxim d'execució de codi",
|
||||
"Code formatted successfully": "Codi formatat correctament",
|
||||
"Code Interpreter": "Intèrpret de codi",
|
||||
"Code Interpreter Engine": "",
|
||||
"Code Interpreter Prompt Template": "",
|
||||
"Code Interpreter Engine": "Motor de l'intèrpret de codi",
|
||||
"Code Interpreter Prompt Template": "Plantilla de la indicació de l'intèrpret de codi",
|
||||
"Collection": "Col·lecció",
|
||||
"Color": "Color",
|
||||
"ComfyUI": "ComfyUI",
|
||||
@ -203,7 +203,7 @@
|
||||
"Confirm Password": "Confirmar la contrasenya",
|
||||
"Confirm your action": "Confirma la teva acció",
|
||||
"Confirm your new password": "Confirma la teva nova contrasenya",
|
||||
"Connect to your own OpenAI compatible API endpoints.": "",
|
||||
"Connect to your own OpenAI compatible API endpoints.": "Connecta als teus propis punts de connexió de l'API compatible amb OpenAI",
|
||||
"Connections": "Connexions",
|
||||
"Constrains effort on reasoning for reasoning models. Only applicable to reasoning models from specific providers that support reasoning effort. (Default: medium)": "Restringeix l'esforç de raonament dels models de raonament. Només aplicable a models de raonament de proveïdors específics que donen suport a l'esforç de raonament. (Per defecte: mitjà)",
|
||||
"Contact Admin for WebUI Access": "Posat en contacte amb l'administrador per accedir a WebUI",
|
||||
@ -215,7 +215,7 @@
|
||||
"Continue with Email": "Continuar amb el correu",
|
||||
"Continue with LDAP": "Continuar amb LDAP",
|
||||
"Control how message text is split for TTS requests. 'Punctuation' splits into sentences, 'paragraphs' splits into paragraphs, and 'none' keeps the message as a single string.": "Controlar com es divideix el text del missatge per a les sol·licituds TTS. 'Puntuació' divideix en frases, 'paràgrafs' divideix en paràgrafs i 'cap' manté el missatge com una cadena única.",
|
||||
"Control the repetition of token sequences in the generated text. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 1.1) will be more lenient. At 1, it is disabled. (Default: 1.1)": "",
|
||||
"Control the repetition of token sequences in the generated text. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 1.1) will be more lenient. At 1, it is disabled. (Default: 1.1)": "Controlar la repetició de seqüències de tokens en el text generat. Un valor més alt (p. ex., 1,5) penalitzarà les repeticions amb més força, mentre que un valor més baix (p. ex., 1,1) serà més indulgent. A l'1, està desactivat. (Per defecte: 1.1)",
|
||||
"Controls": "Controls",
|
||||
"Controls the balance between coherence and diversity of the output. A lower value will result in more focused and coherent text. (Default: 5.0)": "Controlar l'equilibri entre la coherència i la diversitat de la sortida. Un valor més baix donarà lloc a un text més enfocat i coherent. (Per defecte: 5.0)",
|
||||
"Copied": "Copiat",
|
||||
@ -227,7 +227,7 @@
|
||||
"Copy Link": "Copiar l'enllaç",
|
||||
"Copy to clipboard": "Copiar al porta-retalls",
|
||||
"Copying to clipboard was successful!": "La còpia al porta-retalls s'ha realitzat correctament",
|
||||
"CORS must be properly configured by the provider to allow requests from Open WebUI.": "",
|
||||
"CORS must be properly configured by the provider to allow requests from Open WebUI.": "CORS ha de ser configurat correctament pel proveïdor per permetre les sol·licituds d'Open WebUI",
|
||||
"Create": "Crear",
|
||||
"Create a knowledge base": "Crear una base de coneixement",
|
||||
"Create a model": "Crear un model",
|
||||
@ -271,7 +271,7 @@
|
||||
"Delete folder?": "Eliminar la carpeta?",
|
||||
"Delete function?": "Eliminar funció?",
|
||||
"Delete Message": "Eleiminar el missatge",
|
||||
"Delete message?": "",
|
||||
"Delete message?": "Eliminar el missatge?",
|
||||
"Delete prompt?": "Eliminar indicació?",
|
||||
"delete this link": "Eliminar aquest enllaç",
|
||||
"Delete tool?": "Eliminar eina?",
|
||||
@ -282,15 +282,15 @@
|
||||
"Describe your knowledge base and objectives": "Descriu la teva base de coneixement i objectius",
|
||||
"Description": "Descripció",
|
||||
"Didn't fully follow instructions": "No s'han seguit les instruccions completament",
|
||||
"Direct Connections": "",
|
||||
"Direct Connections allow users to connect to their own OpenAI compatible API endpoints.": "",
|
||||
"Direct Connections settings updated": "",
|
||||
"Direct Connections": "Connexions directes",
|
||||
"Direct Connections allow users to connect to their own OpenAI compatible API endpoints.": "Les connexions directes permeten als usuaris connectar-se als seus propis endpoints d'API compatibles amb OpenAI.",
|
||||
"Direct Connections settings updated": "Configuració de les connexions directes actualitzada",
|
||||
"Disabled": "Deshabilitat",
|
||||
"Discover a function": "Descobrir una funció",
|
||||
"Discover a model": "Descobrir un model",
|
||||
"Discover a prompt": "Descobrir una indicació",
|
||||
"Discover a tool": "Descobrir una eina",
|
||||
"Discover how to use Open WebUI and seek support from the community.": "",
|
||||
"Discover how to use Open WebUI and seek support from the community.": "Descobreix com utilitzar Open WebUI i demana suport a la comunitat.",
|
||||
"Discover wonders": "Descobrir meravelles",
|
||||
"Discover, download, and explore custom functions": "Descobrir, descarregar i explorar funcions personalitzades",
|
||||
"Discover, download, and explore custom prompts": "Descobrir, descarregar i explorar indicacions personalitzades",
|
||||
@ -315,14 +315,14 @@
|
||||
"Don't like the style": "No t'agrada l'estil?",
|
||||
"Done": "Fet",
|
||||
"Download": "Descarregar",
|
||||
"Download as SVG": "",
|
||||
"Download as SVG": "Descarrega com a SVG",
|
||||
"Download canceled": "Descàrrega cancel·lada",
|
||||
"Download Database": "Descarregar la base de dades",
|
||||
"Drag and drop a file to upload or select a file to view": "Arrossegar un arxiu per pujar o escull un arxiu a veure",
|
||||
"Draw": "Dibuixar",
|
||||
"Drop any files here to add to the conversation": "Deixa qualsevol arxiu aquí per afegir-lo a la conversa",
|
||||
"e.g. '30s','10m'. Valid time units are 's', 'm', 'h'.": "p. ex. '30s','10m'. Les unitats de temps vàlides són 's', 'm', 'h'.",
|
||||
"e.g. 60": "",
|
||||
"e.g. 60": "p. ex. 60",
|
||||
"e.g. A filter to remove profanity from text": "p. ex. Un filtre per eliminar paraules malsonants del text",
|
||||
"e.g. My Filter": "p. ex. El meu filtre",
|
||||
"e.g. My Tools": "p. ex. Les meves eines",
|
||||
@ -346,7 +346,7 @@
|
||||
"Embedding model set to \"{{embedding_model}}\"": "Model d'incrustació configurat a \"{{embedding_model}}\"",
|
||||
"Enable API Key": "Activar la Clau API",
|
||||
"Enable autocomplete generation for chat messages": "Activar la generació automàtica per als missatges del xat",
|
||||
"Enable Code Interpreter": "",
|
||||
"Enable Code Interpreter": "Activar l'intèrpret de codi",
|
||||
"Enable Community Sharing": "Activar l'ús compartit amb la comunitat",
|
||||
"Enable Google Drive": "Activar Google Drive",
|
||||
"Enable Memory Locking (mlock) to prevent model data from being swapped out of RAM. This option locks the model's working set of pages into RAM, ensuring that they will not be swapped out to disk. This can help maintain performance by avoiding page faults and ensuring fast data access.": "Activar el bloqueig de memòria (mlock) per evitar que les dades del model s'intercanviïn fora de la memòria RAM. Aquesta opció bloqueja el conjunt de pàgines de treball del model a la memòria RAM, assegurant-se que no s'intercanviaran al disc. Això pot ajudar a mantenir el rendiment evitant errors de pàgina i garantint un accés ràpid a les dades.",
|
||||
@ -365,7 +365,7 @@
|
||||
"Enter Application DN Password": "Introdueix la contrasenya del DN d'aplicació",
|
||||
"Enter Bing Search V7 Endpoint": "Introdueix el punt de connexió de Bing Search V7",
|
||||
"Enter Bing Search V7 Subscription Key": "Introdueix la clau de subscripció de Bing Search V7",
|
||||
"Enter Bocha Search API Key": "",
|
||||
"Enter Bocha Search API Key": "Introdueix la clau API de Bocha Search",
|
||||
"Enter Brave Search API Key": "Introdueix la clau API de Brave Search",
|
||||
"Enter certificate path": "Introdueix el camí del certificat",
|
||||
"Enter CFG Scale (e.g. 7.0)": "Entra l'escala CFG (p.ex. 7.0)",
|
||||
@ -379,9 +379,9 @@
|
||||
"Enter Google PSE Engine Id": "Introdueix l'identificador del motor PSE de Google",
|
||||
"Enter Image Size (e.g. 512x512)": "Introdueix la mida de la imatge (p. ex. 512x512)",
|
||||
"Enter Jina API Key": "Introdueix la clau API de Jina",
|
||||
"Enter Jupyter Password": "",
|
||||
"Enter Jupyter Token": "",
|
||||
"Enter Jupyter URL": "",
|
||||
"Enter Jupyter Password": "Introdueix la contrasenya de Jupyter",
|
||||
"Enter Jupyter Token": "Introdueix el token de Jupyter",
|
||||
"Enter Jupyter URL": "Introdueix la URL de Jupyter",
|
||||
"Enter Kagi Search API Key": "Introdueix la clau API de Kagi Search",
|
||||
"Enter language codes": "Introdueix els codis de llenguatge",
|
||||
"Enter Model ID": "Introdueix l'identificador del model",
|
||||
@ -397,8 +397,8 @@
|
||||
"Enter SearchApi Engine": "Introdueix el motor SearchApi",
|
||||
"Enter Searxng Query URL": "Introdueix l'URL de consulta de Searxng",
|
||||
"Enter Seed": "Introdueix la llavor",
|
||||
"Enter SerpApi API Key": "",
|
||||
"Enter SerpApi Engine": "",
|
||||
"Enter SerpApi API Key": "Introdueix la clau API SerpApi",
|
||||
"Enter SerpApi Engine": "Introdueix el motor API SerpApi",
|
||||
"Enter Serper API Key": "Introdueix la clau API Serper",
|
||||
"Enter Serply API Key": "Introdueix la clau API Serply",
|
||||
"Enter Serpstack API Key": "Introdueix la clau API Serpstack",
|
||||
@ -410,7 +410,7 @@
|
||||
"Enter Tavily API Key": "Introdueix la clau API de Tavily",
|
||||
"Enter the public URL of your WebUI. This URL will be used to generate links in the notifications.": "Entra la URL pública de WebUI. Aquesta URL s'utilitzarà per generar els enllaços en les notificacions.",
|
||||
"Enter Tika Server URL": "Introdueix l'URL del servidor Tika",
|
||||
"Enter timeout in seconds": "",
|
||||
"Enter timeout in seconds": "Entra el temps màxim en segons",
|
||||
"Enter Top K": "Introdueix Top K",
|
||||
"Enter URL (e.g. http://127.0.0.1:7860/)": "Introdueix l'URL (p. ex. http://127.0.0.1:7860/)",
|
||||
"Enter URL (e.g. http://localhost:11434)": "Introdueix l'URL (p. ex. http://localhost:11434)",
|
||||
@ -458,7 +458,7 @@
|
||||
"Failed to save models configuration": "No s'ha pogut desar la configuració dels models",
|
||||
"Failed to update settings": "No s'han pogut actualitzar les preferències",
|
||||
"Failed to upload file.": "No s'ha pogut pujar l'arxiu.",
|
||||
"Features": "",
|
||||
"Features": "Característiques",
|
||||
"Features Permissions": "Permisos de les característiques",
|
||||
"February": "Febrer",
|
||||
"Feedback History": "Històric de comentaris",
|
||||
@ -488,7 +488,7 @@
|
||||
"Form": "Formulari",
|
||||
"Format your variables using brackets like this:": "Formata les teves variables utilitzant claudàtors així:",
|
||||
"Frequency Penalty": "Penalització per freqüència",
|
||||
"Full Context Mode": "",
|
||||
"Full Context Mode": "Mode de context complert",
|
||||
"Function": "Funció",
|
||||
"Function Calling": "Crida a funcions",
|
||||
"Function created successfully": "La funció s'ha creat correctament",
|
||||
@ -503,9 +503,9 @@
|
||||
"Functions allow arbitrary code execution": "Les funcions permeten l'execució de codi arbitrari",
|
||||
"Functions allow arbitrary code execution.": "Les funcions permeten l'execució de codi arbitrari.",
|
||||
"Functions imported successfully": "Les funcions s'han importat correctament",
|
||||
"Gemini": "",
|
||||
"Gemini API Config": "",
|
||||
"Gemini API Key is required.": "",
|
||||
"Gemini": "Gemini",
|
||||
"Gemini API Config": "Configuració de Gemini API",
|
||||
"Gemini API Key is required.": "La clau API de Gemini és necessària",
|
||||
"General": "General",
|
||||
"General Settings": "Preferències generals",
|
||||
"Generate an image": "Generar una imatge",
|
||||
@ -532,7 +532,7 @@
|
||||
"Hex Color": "Color hexadecimal",
|
||||
"Hex Color - Leave empty for default color": "Color hexadecimal - Deixar buit per a color per defecte",
|
||||
"Hide": "Amaga",
|
||||
"Home": "",
|
||||
"Home": "Inici",
|
||||
"Host": "Servidor",
|
||||
"How can I help you today?": "Com et puc ajudar avui?",
|
||||
"How would you rate this response?": "Com avaluaries aquesta resposta?",
|
||||
@ -576,8 +576,8 @@
|
||||
"JSON Preview": "Vista prèvia del document JSON",
|
||||
"July": "Juliol",
|
||||
"June": "Juny",
|
||||
"Jupyter Auth": "",
|
||||
"Jupyter URL": "",
|
||||
"Jupyter Auth": "Autenticació Jupyter",
|
||||
"Jupyter URL": "URL de Jupyter",
|
||||
"JWT Expiration": "Caducitat del JWT",
|
||||
"JWT Token": "Token JWT",
|
||||
"Kagi Search API Key": "Clau API de Kagi Search",
|
||||
@ -607,12 +607,12 @@
|
||||
"Leave empty to include all models or select specific models": "Deixa-ho en blanc per incloure tots els models o selecciona models específics",
|
||||
"Leave empty to use the default prompt, or enter a custom prompt": "Deixa-ho en blanc per utilitzar la indicació predeterminada o introdueix una indicació personalitzada",
|
||||
"Leave model field empty to use the default model.": "Deixa el camp de model buit per utilitzar el model per defecte.",
|
||||
"License": "",
|
||||
"License": "Llicència",
|
||||
"Light": "Clar",
|
||||
"Listening...": "Escoltant...",
|
||||
"Llama.cpp": "Llama.cpp",
|
||||
"LLMs can make mistakes. Verify important information.": "Els models de llenguatge poden cometre errors. Verifica la informació important.",
|
||||
"Loading Kokoro.js...": "",
|
||||
"Loading Kokoro.js...": "Carregant Kokoro.js",
|
||||
"Local": "Local",
|
||||
"Local Models": "Models locals",
|
||||
"Lost": "Perdut",
|
||||
@ -622,7 +622,7 @@
|
||||
"Make sure to export a workflow.json file as API format from ComfyUI.": "Assegura't d'exportar un fitxer workflow.json com a format API des de ComfyUI.",
|
||||
"Manage": "Gestionar",
|
||||
"Manage Arena Models": "Gestionar els models de l'Arena",
|
||||
"Manage Direct Connections": "",
|
||||
"Manage Direct Connections": "Gestionar les connexions directes",
|
||||
"Manage Models": "Gestionar els models",
|
||||
"Manage Ollama": "Gestionar Ollama",
|
||||
"Manage Ollama API Connections": "Gestionar les connexions a l'API d'Ollama",
|
||||
@ -766,7 +766,7 @@
|
||||
"Plain text (.txt)": "Text pla (.txt)",
|
||||
"Playground": "Zona de jocs",
|
||||
"Please carefully review the following warnings:": "Si us plau, revisa els següents avisos amb cura:",
|
||||
"Please do not close the settings page while loading the model.": "",
|
||||
"Please do not close the settings page while loading the model.": "No tanquis la pàgina de configuració mentre carregues el model.",
|
||||
"Please enter a prompt": "Si us plau, entra una indicació",
|
||||
"Please fill in all fields.": "Emplena tots els camps, si us plau.",
|
||||
"Please select a model first.": "Si us plau, selecciona un model primer",
|
||||
@ -776,7 +776,7 @@
|
||||
"Positive attitude": "Actitud positiva",
|
||||
"Prefix ID": "Identificador del prefix",
|
||||
"Prefix ID is used to avoid conflicts with other connections by adding a prefix to the model IDs - leave empty to disable": "L'identificador de prefix s'utilitza per evitar conflictes amb altres connexions afegint un prefix als ID de model; deixa'l en blanc per desactivar-lo.",
|
||||
"Presence Penalty": "",
|
||||
"Presence Penalty": "Penalització de presència",
|
||||
"Previous 30 days": "30 dies anteriors",
|
||||
"Previous 7 days": "7 dies anteriors",
|
||||
"Profile Image": "Imatge de perfil",
|
||||
@ -813,7 +813,7 @@
|
||||
"Rename": "Canviar el nom",
|
||||
"Reorder Models": "Reordenar els models",
|
||||
"Repeat Last N": "Repeteix els darrers N",
|
||||
"Repeat Penalty (Ollama)": "",
|
||||
"Repeat Penalty (Ollama)": "Penalització per repetició (Ollama)",
|
||||
"Reply in Thread": "Respondre al fil",
|
||||
"Request Mode": "Mode de sol·licitud",
|
||||
"Reranking Model": "Model de reavaluació",
|
||||
@ -876,7 +876,7 @@
|
||||
"Select a pipeline": "Seleccionar una Pipeline",
|
||||
"Select a pipeline url": "Seleccionar l'URL d'una Pipeline",
|
||||
"Select a tool": "Seleccionar una eina",
|
||||
"Select an auth method": "",
|
||||
"Select an auth method": "Seleccionar un mètode d'autenticació",
|
||||
"Select an Ollama instance": "Seleccionar una instància d'Ollama",
|
||||
"Select Engine": "Seleccionar el motor",
|
||||
"Select Knowledge": "Seleccionar coneixement",
|
||||
@ -889,8 +889,8 @@
|
||||
"Send message": "Enviar missatge",
|
||||
"Sends `stream_options: { include_usage: true }` in the request.\nSupported providers will return token usage information in the response when set.": "Envia `stream_options: { include_usage: true }` a la sol·licitud.\nEls proveïdors compatibles retornaran la informació d'ús del token a la resposta quan s'estableixi.",
|
||||
"September": "Setembre",
|
||||
"SerpApi API Key": "",
|
||||
"SerpApi Engine": "",
|
||||
"SerpApi API Key": "Clau API de SerpApi",
|
||||
"SerpApi Engine": "Motor de SerpApi",
|
||||
"Serper API Key": "Clau API de Serper",
|
||||
"Serply API Key": "Clau API de Serply",
|
||||
"Serpstack API Key": "Clau API de Serpstack",
|
||||
@ -910,8 +910,8 @@
|
||||
"Set the number of worker threads used for computation. This option controls how many threads are used to process incoming requests concurrently. Increasing this value can improve performance under high concurrency workloads but may also consume more CPU resources.": "Establir el nombre de fils de treball utilitzats per al càlcul. Aquesta opció controla quants fils s'utilitzen per processar les sol·licituds entrants simultàniament. Augmentar aquest valor pot millorar el rendiment amb càrregues de treball de concurrència elevada, però també pot consumir més recursos de CPU.",
|
||||
"Set Voice": "Establir la veu",
|
||||
"Set whisper model": "Establir el model whisper",
|
||||
"Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 0)": "",
|
||||
"Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 1.1)": "",
|
||||
"Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 0)": "Estableix un biaix pla contra tokens que han aparegut almenys una vegada. Un valor més alt (p. ex., 1,5) penalitzarà les repeticions amb més força, mentre que un valor més baix (p. ex., 0,9) serà més indulgent. A 0, està desactivat. (Per defecte: 0)",
|
||||
"Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 1.1)": "Estableix un biaix d'escala contra tokens per penalitzar les repeticions, en funció de quantes vegades han aparegut. Un valor més alt (p. ex., 1,5) penalitzarà les repeticions amb més força, mentre que un valor més baix (p. ex., 0,9) serà més indulgent. A 0, està desactivat. (Per defecte: 1.1)",
|
||||
"Sets how far back for the model to look back to prevent repetition. (Default: 64, 0 = disabled, -1 = num_ctx)": "Establir fins a quin punt el model mira enrere per evitar la repetició. (Per defecte: 64, 0 = desactivat, -1 = num_ctx)",
|
||||
"Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt. (Default: random)": "Establir la llavor del nombre aleatori que s'utilitzarà per a la generació. Establir-ho a un número específic farà que el model generi el mateix text per a la mateixa sol·licitud. (Per defecte: aleatori)",
|
||||
"Sets the size of the context window used to generate the next token. (Default: 2048)": "Estableix la mida de la finestra de context utilitzada per generar el següent token. (Per defecte: 2048)",
|
||||
@ -958,7 +958,7 @@
|
||||
"Tags Generation Prompt": "Indicació per a la generació d'etiquetes",
|
||||
"Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1)": "El mostreig sense cua s'utilitza per reduir l'impacte de tokens menys probables de la sortida. Un valor més alt (p. ex., 2,0) reduirà més l'impacte, mentre que un valor d'1,0 desactiva aquesta configuració. (per defecte: 1)",
|
||||
"Tap to interrupt": "Prem per interrompre",
|
||||
"Tasks": "",
|
||||
"Tasks": "Tasques",
|
||||
"Tavily API Key": "Clau API de Tavily",
|
||||
"Tell us more:": "Dona'ns més informació:",
|
||||
"Temperature": "Temperatura",
|
||||
@ -1005,7 +1005,7 @@
|
||||
"Title (e.g. Tell me a fun fact)": "Títol (p. ex. Digues-me quelcom divertit)",
|
||||
"Title Auto-Generation": "Generació automàtica de títol",
|
||||
"Title cannot be an empty string.": "El títol no pot ser una cadena buida.",
|
||||
"Title Generation": "",
|
||||
"Title Generation": "Generació de títols",
|
||||
"Title Generation Prompt": "Indicació de generació de títol",
|
||||
"TLS": "TLS",
|
||||
"To access the available model names for downloading,": "Per accedir als noms dels models disponibles per descarregar,",
|
||||
@ -1062,7 +1062,7 @@
|
||||
"Updated": "Actualitzat",
|
||||
"Updated at": "Actualitzat el",
|
||||
"Updated At": "Actualitzat el",
|
||||
"Upgrade to a licensed plan for enhanced capabilities, including custom theming and branding, and dedicated support.": "",
|
||||
"Upgrade to a licensed plan for enhanced capabilities, including custom theming and branding, and dedicated support.": "Actualitzar a un pla amb llicència per obtenir capacitats millorades, com ara la temàtica personalitzada i la marca, i assistència dedicada.",
|
||||
"Upload": "Pujar",
|
||||
"Upload a GGUF model": "Pujar un model GGUF",
|
||||
"Upload directory": "Pujar directori",
|
||||
@ -1101,7 +1101,7 @@
|
||||
"Warning:": "Avís:",
|
||||
"Warning: Enabling this will allow users to upload arbitrary code on the server.": "Avís: Habilitar això permetrà als usuaris penjar codi arbitrari al servidor.",
|
||||
"Warning: If you update or change your embedding model, you will need to re-import all documents.": "Avís: Si s'actualitza o es canvia el model d'incrustació, s'hauran de tornar a importar tots els documents.",
|
||||
"Warning: Jupyter execution enables arbitrary code execution, posing severe security risks—proceed with extreme caution.": "",
|
||||
"Warning: Jupyter execution enables arbitrary code execution, posing severe security risks—proceed with extreme caution.": "Avís: l'execució de Jupyter permet l'execució de codi arbitrari, la qual cosa comporta greus riscos de seguretat; procediu amb extrema precaució.",
|
||||
"Web": "Web",
|
||||
"Web API": "Web API",
|
||||
"Web Loader Settings": "Preferències del carregador web",
|
||||
|
@ -1,5 +1,5 @@
|
||||
{
|
||||
"-1 for no limit, or a positive integer for a specific limit": "-1 表示无限制,正整数表示具体限制”",
|
||||
"-1 for no limit, or a positive integer for a specific limit": "-1 表示无限制,正整数表示具体限制",
|
||||
"'s', 'm', 'h', 'd', 'w' or '-1' for no expiration.": "'s', 'm', 'h', 'd', 'w' 或 '-1' 表示无过期时间。",
|
||||
"(e.g. `sh webui.sh --api --api-auth username_password`)": "(例如 `sh webui.sh --api --api-auth username_password`)",
|
||||
"(e.g. `sh webui.sh --api`)": "(例如 `sh webui.sh --api`)",
|
||||
@ -63,7 +63,7 @@
|
||||
"Allow Voice Interruption in Call": "允许通话中的打断语音",
|
||||
"Allowed Endpoints": "允许的端点",
|
||||
"Already have an account?": "已经拥有账号了?",
|
||||
"Alternative to the top_p, and aims to ensure a balance of quality and variety. The parameter p represents the minimum probability for a token to be considered, relative to the probability of the most likely token. For example, with p=0.05 and the most likely token having a probability of 0.9, logits with a value less than 0.045 are filtered out. (Default: 0.0)": "top_p的替代方法,目标是在质量和多样性之间取得平衡。参数p表示一个token相对于最有可能的token所需的最低概率。比如,当p=0.05且最有可能的token概率为0.9时,概率低于0.045的logits会被排除。(默认值:0.0)",
|
||||
"Alternative to the top_p, and aims to ensure a balance of quality and variety. The parameter p represents the minimum probability for a token to be considered, relative to the probability of the most likely token. For example, with p=0.05 and the most likely token having a probability of 0.9, logits with a value less than 0.045 are filtered out. (Default: 0.0)": "top_p的替代方法,目标是在质量和多样性之间取得平衡。参数p表示一个Token相对于最有可能的Token所需的最低概率。比如,当p=0.05且最有可能的Token概率为0.9时,概率低于0.045的logits会被排除。(默认值:0.0)",
|
||||
"Always": "保持",
|
||||
"Amazing": "很棒",
|
||||
"an assistant": "一个助手",
|
||||
@ -182,7 +182,7 @@
|
||||
"Code execution": "代码执行",
|
||||
"Code Execution": "代码执行",
|
||||
"Code Execution Engine": "代码执行引擎",
|
||||
"Code Execution Timeout": "",
|
||||
"Code Execution Timeout": "代码执行超时时间",
|
||||
"Code formatted successfully": "代码格式化成功",
|
||||
"Code Interpreter": "代码解释器",
|
||||
"Code Interpreter Engine": "代码解释引擎",
|
||||
@ -205,7 +205,7 @@
|
||||
"Confirm your new password": "确认新密码",
|
||||
"Connect to your own OpenAI compatible API endpoints.": "连接到你自己的与 OpenAI 兼容的 API 接口端点。",
|
||||
"Connections": "外部连接",
|
||||
"Constrains effort on reasoning for reasoning models. Only applicable to reasoning models from specific providers that support reasoning effort. (Default: medium)": "限制推理模型的推理努力。仅适用于支持推理努力的特定提供商的推理模型。(默认值:中等)",
|
||||
"Constrains effort on reasoning for reasoning models. Only applicable to reasoning models from specific providers that support reasoning effort. (Default: medium)": "限制模型的努力。仅适用于支持努力的特定提供商的模型。(默认值:中等)",
|
||||
"Contact Admin for WebUI Access": "请联系管理员以获取访问权限",
|
||||
"Content": "内容",
|
||||
"Content Extraction": "内容提取",
|
||||
@ -322,7 +322,7 @@
|
||||
"Draw": "平局",
|
||||
"Drop any files here to add to the conversation": "拖动文件到此处以添加到对话中",
|
||||
"e.g. '30s','10m'. Valid time units are 's', 'm', 'h'.": "例如 '30s','10m'。有效的时间单位是秒:'s',分:'m',时:'h'。",
|
||||
"e.g. 60": "",
|
||||
"e.g. 60": "例如 '60'",
|
||||
"e.g. A filter to remove profanity from text": "例如:一个用于过滤文本中不当内容的过滤器",
|
||||
"e.g. My Filter": "例如:我的过滤器",
|
||||
"e.g. My Tools": "例如:我的工具",
|
||||
@ -380,7 +380,7 @@
|
||||
"Enter Image Size (e.g. 512x512)": "输入图像分辨率 (例如:512x512)",
|
||||
"Enter Jina API Key": "输入 Jina API 密钥",
|
||||
"Enter Jupyter Password": "输入 Jupyter 密码",
|
||||
"Enter Jupyter Token": "输入 Jupyter Token",
|
||||
"Enter Jupyter Token": "输入 Jupyter 令牌",
|
||||
"Enter Jupyter URL": "输入 Jupyter URL",
|
||||
"Enter Kagi Search API Key": "输入 Kagi Search API 密钥",
|
||||
"Enter language codes": "输入语言代码",
|
||||
@ -410,7 +410,7 @@
|
||||
"Enter Tavily API Key": "输入 Tavily API 密钥",
|
||||
"Enter the public URL of your WebUI. This URL will be used to generate links in the notifications.": "输入 WebUI 的公共 URL。此 URL 将用于在通知中生成链接。",
|
||||
"Enter Tika Server URL": "输入 Tika 服务器地址",
|
||||
"Enter timeout in seconds": "",
|
||||
"Enter timeout in seconds": "输入以秒为单位的超时时间",
|
||||
"Enter Top K": "输入 Top K",
|
||||
"Enter URL (e.g. http://127.0.0.1:7860/)": "输入地址 (例如:http://127.0.0.1:7860/)",
|
||||
"Enter URL (e.g. http://localhost:11434)": "输入地址 (例如:http://localhost:11434)",
|
||||
@ -503,9 +503,9 @@
|
||||
"Functions allow arbitrary code execution": "注意:函数有权执行任意代码",
|
||||
"Functions allow arbitrary code execution.": "注意:函数有权执行任意代码。",
|
||||
"Functions imported successfully": "函数导入成功",
|
||||
"Gemini": "",
|
||||
"Gemini API Config": "",
|
||||
"Gemini API Key is required.": "",
|
||||
"Gemini": "Gemini",
|
||||
"Gemini API Config": "Gemini API 配置",
|
||||
"Gemini API Key is required.": "需要 Gemini API 密钥。",
|
||||
"General": "通用",
|
||||
"General Settings": "通用设置",
|
||||
"Generate an image": "生成图像",
|
||||
@ -629,7 +629,7 @@
|
||||
"Manage OpenAI API Connections": "管理OpenAI API连接",
|
||||
"Manage Pipelines": "管理 Pipeline",
|
||||
"March": "三月",
|
||||
"Max Tokens (num_predict)": "最多 Token (num_predict)",
|
||||
"Max Tokens (num_predict)": "最大Token数量 (num_predict)",
|
||||
"Max Upload Count": "最大上传数量",
|
||||
"Max Upload Size": "最大上传大小",
|
||||
"Maximum of 3 models can be downloaded simultaneously. Please try again later.": "最多可以同时下载 3 个模型,请稍后重试。",
|
||||
@ -910,14 +910,14 @@
|
||||
"Set the number of worker threads used for computation. This option controls how many threads are used to process incoming requests concurrently. Increasing this value can improve performance under high concurrency workloads but may also consume more CPU resources.": "设置用于计算的工作线程数量。该选项可控制并发处理传入请求的线程数量。增加该值可以提高高并发工作负载下的性能,但也可能消耗更多的 CPU 资源。",
|
||||
"Set Voice": "设置音色",
|
||||
"Set whisper model": "设置 whisper 模型",
|
||||
"Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 0)": "这个设置项用于调整对重复 tokens 的抑制强度。当某个 token 至少出现过一次后,系统会通过 flat bias 参数施加惩罚力度:数值越大(如 1.5),抑制重复的效果越强烈;数值较小(如 0.9)则相对宽容。当设为 0 时,系统会完全关闭这个重复抑制功能(默认值为 0)。",
|
||||
"Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 1.1)": "这个参数用于通过 scaling bias 机制抑制重复内容:当某些 tokens 重复出现时,系统会根据它们已出现的次数自动施加惩罚。数值越大(如 1.5)惩罚力度越强,能更有效减少重复;数值较小(如 0.9)则允许更多重复。当设为 0 时完全关闭该功能,默认值设置为 1.1 保持适度抑制。",
|
||||
"Sets a flat bias against tokens that have appeared at least once. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 0)": "这个设置项用于调整对重复 Token 的抑制强度。当某个 Token 至少出现过一次后,系统会通过 flat bias 参数施加惩罚力度:数值越大(如 1.5),抑制重复的效果越强烈;数值较小(如 0.9)则相对宽容。当设为 0 时,系统会完全关闭这个重复抑制功能(默认值为 0)。",
|
||||
"Sets a scaling bias against tokens to penalize repetitions, based on how many times they have appeared. A higher value (e.g., 1.5) will penalize repetitions more strongly, while a lower value (e.g., 0.9) will be more lenient. At 0, it is disabled. (Default: 1.1)": "这个参数用于通过 scaling bias 机制抑制重复内容:当某些 Token 重复出现时,系统会根据它们已出现的次数自动施加惩罚。数值越大(如 1.5)惩罚力度越强,能更有效减少重复;数值较小(如 0.9)则允许更多重复。当设为 0 时完全关闭该功能,默认值设置为 1.1 保持适度抑制。",
|
||||
"Sets how far back for the model to look back to prevent repetition. (Default: 64, 0 = disabled, -1 = num_ctx)": "设置模型回溯多远以防止重复。(默认值:64,0 = 禁用,-1 = num_ctx)",
|
||||
"Sets the random number seed to use for generation. Setting this to a specific number will make the model generate the same text for the same prompt. (Default: random)": "设置 random number seed 可以控制模型生成文本的随机起点。如果指定一个具体数字,当输入相同的提示语时,模型每次都会生成完全相同的文本内容(默认是随机选取 seed)。",
|
||||
"Sets the size of the context window used to generate the next token. (Default: 2048)": "设置用于生成下一个 Token 的上下文大小。(默认值:2048)",
|
||||
"Sets the stop sequences to use. When this pattern is encountered, the LLM will stop generating text and return. Multiple stop patterns may be set by specifying multiple separate stop parameters in a modelfile.": "设置要使用的停止序列。遇到这种模式时,大语言模型将停止生成文本并返回。可以通过在模型文件中指定多个单独的停止参数来设置多个停止模式。",
|
||||
"Settings": "设置",
|
||||
"Settings saved successfully!": "设置已保存",
|
||||
"Settings saved successfully!": "设置已成功保存!",
|
||||
"Share": "分享",
|
||||
"Share Chat": "分享对话",
|
||||
"Share to Open WebUI Community": "分享到 OpenWebUI 社区",
|
||||
@ -956,7 +956,7 @@
|
||||
"System Prompt": "系统提示词 (System Prompt)",
|
||||
"Tags Generation": "标签生成",
|
||||
"Tags Generation Prompt": "标签生成提示词",
|
||||
"Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1)": "Tail free sampling 用于减少输出中可能性较低的标记的影响。数值越大(如 2.0),影响就越小,而数值为 1.0 则会禁用此设置。(默认值:1)",
|
||||
"Tail free sampling is used to reduce the impact of less probable tokens from the output. A higher value (e.g., 2.0) will reduce the impact more, while a value of 1.0 disables this setting. (default: 1)": "Tail free sampling 用于减少输出中可能性较低的Token的影响。数值越大(如 2.0),影响就越小,而数值为 1.0 则会禁用此设置。(默认值:1)",
|
||||
"Tap to interrupt": "点击以中断",
|
||||
"Tasks": "任务",
|
||||
"Tavily API Key": "Tavily API 密钥",
|
||||
@ -981,11 +981,11 @@
|
||||
"The score should be a value between 0.0 (0%) and 1.0 (100%).": "分值应介于 0.0(0%)和 1.0(100%)之间。",
|
||||
"The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8)": "模型的温度。提高温度将使模型更具创造性地回答。(默认值:0.8)",
|
||||
"Theme": "主题",
|
||||
"Thinking...": "正在深度思考...",
|
||||
"Thinking...": "正在思考...",
|
||||
"This action cannot be undone. Do you wish to continue?": "此操作无法撤销。是否确认继续?",
|
||||
"This ensures that your valuable conversations are securely saved to your backend database. Thank you!": "这将确保您的宝贵对话被安全地保存到后台数据库中。感谢!",
|
||||
"This is an experimental feature, it may not function as expected and is subject to change at any time.": "这是一个实验功能,可能不会如预期那样工作,而且可能随时发生变化。",
|
||||
"This option controls how many tokens are preserved when refreshing the context. For example, if set to 2, the last 2 tokens of the conversation context will be retained. Preserving context can help maintain the continuity of a conversation, but it may reduce the ability to respond to new topics. (Default: 24)": "该选项控制刷新上下文时保留多少标记。例如,如果设置为 2,就会保留对话上下文的最后 2 个标记。保留上下文有助于保持对话的连续性,但可能会降低回复新话题的能力。(默认值:24)",
|
||||
"This option controls how many tokens are preserved when refreshing the context. For example, if set to 2, the last 2 tokens of the conversation context will be retained. Preserving context can help maintain the continuity of a conversation, but it may reduce the ability to respond to new topics. (Default: 24)": "该选项控制刷新上下文时保留多少Token。例如,如果设置为 2,就会保留对话上下文的最后 2 个Token。保留上下文有助于保持对话的连续性,但可能会降低回复新话题的能力。(默认值:24)",
|
||||
"This option sets the maximum number of tokens the model can generate in its response. Increasing this limit allows the model to provide longer answers, but it may also increase the likelihood of unhelpful or irrelevant content being generated. (Default: 128)": "此选项设置了模型在回答中可以生成的最大 Token 数。增加这个限制可以让模型提供更长的答案,但也可能增加生成无用或不相关内容的可能性。 (默认值:128)",
|
||||
"This option will delete all existing files in the collection and replace them with newly uploaded files.": "此选项将会删除文件集中所有文件,并用新上传的文件替换。",
|
||||
"This response was generated by \"{{model}}\"": "此回复由 \"{{model}}\" 生成",
|
||||
@ -995,8 +995,8 @@
|
||||
"This will delete all models including custom models and cannot be undone.": "这将删除所有模型,包括自定义模型,且无法撤销。",
|
||||
"This will reset the knowledge base and sync all files. Do you wish to continue?": "这将重置知识库并替换所有文件为目录下文件。确认继续?",
|
||||
"Thorough explanation": "解释较为详细",
|
||||
"Thought for {{DURATION}}": "已深度思考 用时 {{DURATION}}",
|
||||
"Thought for {{DURATION}} seconds": "已深度思考 用时 {{DURATION}} 秒",
|
||||
"Thought for {{DURATION}}": "已推理 持续 {{DURATION}}",
|
||||
"Thought for {{DURATION}} seconds": "已推理 持续 {{DURATION}} 秒",
|
||||
"Tika": "Tika",
|
||||
"Tika Server URL required.": "请输入 Tika 服务器地址。",
|
||||
"Tiktoken": "Tiktoken",
|
||||
@ -1007,7 +1007,7 @@
|
||||
"Title cannot be an empty string.": "标题不能为空。",
|
||||
"Title Generation": "标题生成",
|
||||
"Title Generation Prompt": "用于自动生成标题的提示词",
|
||||
"TLS": "TLS",
|
||||
"TLS": "传输层安全协议",
|
||||
"To access the available model names for downloading,": "要访问可下载的模型名称,",
|
||||
"To access the GGUF models available for downloading,": "要访问可下载的 GGUF 模型,",
|
||||
"To access the WebUI, please reach out to the administrator. Admins can manage user statuses from the Admin Panel.": "请联系管理员以访问。管理员可以在后台管理面板中管理用户状态。",
|
||||
@ -1022,7 +1022,7 @@
|
||||
"Toggle settings": "切换设置",
|
||||
"Toggle sidebar": "切换侧边栏",
|
||||
"Token": "Token",
|
||||
"Tokens To Keep On Context Refresh (num_keep)": "在语境刷新时需保留的 Tokens",
|
||||
"Tokens To Keep On Context Refresh (num_keep)": "在语境刷新时需保留的 Token 数量",
|
||||
"Too verbose": "过于冗长",
|
||||
"Tool created successfully": "工具创建成功",
|
||||
"Tool deleted successfully": "工具删除成功",
|
||||
|
Loading…
x
Reference in New Issue
Block a user