mirror of
https://github.com/ItzCrazyKns/Perplexica.git
synced 2025-04-30 16:22:29 +00:00
64 lines
1.6 KiB
TypeScript
64 lines
1.6 KiB
TypeScript
import { OllamaEmbeddings } from '@langchain/community/embeddings/ollama';
|
|
import { getOllamaApiEndpoint } from '../../config';
|
|
import logger from '../../utils/logger';
|
|
import { ChatOllama } from '@langchain/community/chat_models/ollama';
|
|
|
|
export const loadOllamaChatModels = async () => {
|
|
const ollamaEndpoint = getOllamaApiEndpoint();
|
|
|
|
if (!ollamaEndpoint) return {};
|
|
|
|
try {
|
|
const response = await fetch(`${ollamaEndpoint}/api/tags`, {
|
|
headers: {
|
|
'Content-Type': 'application/json',
|
|
},
|
|
});
|
|
|
|
const { models: ollamaModels } = (await response.json()) as any;
|
|
|
|
const chatModels = ollamaModels.reduce((acc, model) => {
|
|
acc[model.model] = new ChatOllama({
|
|
baseUrl: ollamaEndpoint,
|
|
model: model.model,
|
|
temperature: 0.7,
|
|
});
|
|
return acc;
|
|
}, {});
|
|
|
|
return chatModels;
|
|
} catch (err) {
|
|
logger.error(`Error loading Ollama models: ${err}`);
|
|
return {};
|
|
}
|
|
};
|
|
|
|
export const loadOllamaEmbeddingsModels = async () => {
|
|
const ollamaEndpoint = getOllamaApiEndpoint();
|
|
|
|
if (!ollamaEndpoint) return {};
|
|
|
|
try {
|
|
const response = await fetch(`${ollamaEndpoint}/api/tags`, {
|
|
headers: {
|
|
'Content-Type': 'application/json',
|
|
},
|
|
});
|
|
|
|
const { models: ollamaModels } = (await response.json()) as any;
|
|
|
|
const embeddingsModels = ollamaModels.reduce((acc, model) => {
|
|
acc[model.model] = new OllamaEmbeddings({
|
|
baseUrl: ollamaEndpoint,
|
|
model: model.model,
|
|
});
|
|
return acc;
|
|
}, {});
|
|
|
|
return embeddingsModels;
|
|
} catch (err) {
|
|
logger.error(`Error loading Ollama embeddings model: ${err}`);
|
|
return {};
|
|
}
|
|
};
|