ваше сообщение коммита
This commit is contained in:
@@ -34,7 +34,7 @@ class AIConfigService {
|
||||
this.defaults = {
|
||||
ollama_base_url: process.env.OLLAMA_BASE_URL || 'http://ollama:11434',
|
||||
ollama_llm_model: process.env.OLLAMA_MODEL || 'qwen2.5:7b',
|
||||
ollama_embedding_model: process.env.OLLAMA_EMBED_MODEL || 'mxbai-embed-large:latest',
|
||||
ollama_embedding_model: process.env.OLLAMA_EMBED_MODEL || process.env.OLLAMA_EMBEDDINGS_MODEL || 'mxbai-embed-large:latest',
|
||||
vector_search_url: process.env.VECTOR_SEARCH_URL || 'http://vector-search:8001',
|
||||
embedding_parameters: {
|
||||
batch_size: 32,
|
||||
|
||||
@@ -44,7 +44,7 @@ async function _updateSyncCache() {
|
||||
syncCache = {
|
||||
baseUrl: process.env.OLLAMA_BASE_URL || 'http://ollama:11434',
|
||||
defaultModel: process.env.OLLAMA_MODEL || 'qwen2.5:7b',
|
||||
embeddingModel: process.env.OLLAMA_EMBED_MODEL || 'mxbai-embed-large:latest'
|
||||
embeddingModel: process.env.OLLAMA_EMBED_MODEL || process.env.OLLAMA_EMBEDDINGS_MODEL || 'mxbai-embed-large:latest'
|
||||
};
|
||||
}
|
||||
}
|
||||
@@ -69,7 +69,7 @@ function _getFromSyncCache(key) {
|
||||
const defaults = {
|
||||
baseUrl: process.env.OLLAMA_BASE_URL || 'http://ollama:11434',
|
||||
defaultModel: process.env.OLLAMA_MODEL || 'qwen2.5:7b',
|
||||
embeddingModel: process.env.OLLAMA_EMBED_MODEL || 'mxbai-embed-large:latest'
|
||||
embeddingModel: process.env.OLLAMA_EMBED_MODEL || process.env.OLLAMA_EMBEDDINGS_MODEL || 'mxbai-embed-large:latest'
|
||||
};
|
||||
|
||||
return defaults[key] || defaults.baseUrl;
|
||||
|
||||
@@ -149,6 +149,7 @@ services:
|
||||
- OLLAMA_BASE_URL=http://ollama:11434
|
||||
- OLLAMA_MODEL=${OLLAMA_MODEL:-qwen2.5:7b}
|
||||
- OLLAMA_EMBEDDINGS_MODEL=${OLLAMA_EMBEDDINGS_MODEL:-mxbai-embed-large:latest}
|
||||
- OLLAMA_EMBED_MODEL=${OLLAMA_EMBEDDINGS_MODEL:-mxbai-embed-large:latest}
|
||||
# FRONTEND_URL настраивается в коде, не через env
|
||||
- VECTOR_SEARCH_URL=http://vector-search:8001
|
||||
- LOG_LEVEL=${LOG_LEVEL:-warn}
|
||||
|
||||
@@ -164,6 +164,8 @@ services:
|
||||
- OLLAMA_BASE_URL=http://dapp-ollama:11434
|
||||
- OLLAMA_MODEL=${OLLAMA_MODEL:-qwen2.5:7b}
|
||||
- OLLAMA_EMBEDDINGS_MODEL=${OLLAMA_EMBEDDINGS_MODEL:-mxbai-embed-large:latest}
|
||||
# Backend читает OLLAMA_EMBED_MODEL, не OLLAMA_EMBEDDINGS_MODEL — передаём то же значение
|
||||
- OLLAMA_EMBED_MODEL=${OLLAMA_EMBEDDINGS_MODEL:-mxbai-embed-large:latest}
|
||||
# 🆕 Исправленный URL для Vector Search
|
||||
- VECTOR_SEARCH_URL=http://dapp-vector-search:8001
|
||||
# Команда запуска для production
|
||||
|
||||
Reference in New Issue
Block a user