| const { ErrorTypes, EModelEndpoint, mapModelToAzureConfig } = require('librechat-data-provider'); |
| const { |
| isEnabled, |
| resolveHeaders, |
| isUserProvided, |
| getOpenAIConfig, |
| getAzureCredentials, |
| } = require('@librechat/api'); |
| const { getUserKeyValues, checkUserKeyExpiry } = require('~/server/services/UserService'); |
| const OpenAIClient = require('~/app/clients/OpenAIClient'); |
|
|
| const initializeClient = async ({ |
| req, |
| res, |
| endpointOption, |
| optionsOnly, |
| overrideEndpoint, |
| overrideModel, |
| }) => { |
| const appConfig = req.config; |
| const { |
| PROXY, |
| OPENAI_API_KEY, |
| AZURE_API_KEY, |
| OPENAI_REVERSE_PROXY, |
| AZURE_OPENAI_BASEURL, |
| OPENAI_SUMMARIZE, |
| DEBUG_OPENAI, |
| } = process.env; |
| const { key: expiresAt } = req.body; |
| const modelName = overrideModel ?? req.body.model; |
| const endpoint = overrideEndpoint ?? req.body.endpoint; |
| const contextStrategy = isEnabled(OPENAI_SUMMARIZE) ? 'summarize' : null; |
|
|
| const credentials = { |
| [EModelEndpoint.openAI]: OPENAI_API_KEY, |
| [EModelEndpoint.azureOpenAI]: AZURE_API_KEY, |
| }; |
|
|
| const baseURLOptions = { |
| [EModelEndpoint.openAI]: OPENAI_REVERSE_PROXY, |
| [EModelEndpoint.azureOpenAI]: AZURE_OPENAI_BASEURL, |
| }; |
|
|
| const userProvidesKey = isUserProvided(credentials[endpoint]); |
| const userProvidesURL = isUserProvided(baseURLOptions[endpoint]); |
|
|
| let userValues = null; |
| if (expiresAt && (userProvidesKey || userProvidesURL)) { |
| checkUserKeyExpiry(expiresAt, endpoint); |
| userValues = await getUserKeyValues({ userId: req.user.id, name: endpoint }); |
| } |
|
|
| let apiKey = userProvidesKey ? userValues?.apiKey : credentials[endpoint]; |
| let baseURL = userProvidesURL ? userValues?.baseURL : baseURLOptions[endpoint]; |
|
|
| let clientOptions = { |
| contextStrategy, |
| proxy: PROXY ?? null, |
| debug: isEnabled(DEBUG_OPENAI), |
| reverseProxyUrl: baseURL ? baseURL : null, |
| ...endpointOption, |
| }; |
|
|
| const isAzureOpenAI = endpoint === EModelEndpoint.azureOpenAI; |
| |
| const azureConfig = isAzureOpenAI && appConfig.endpoints?.[EModelEndpoint.azureOpenAI]; |
| let serverless = false; |
| if (isAzureOpenAI && azureConfig) { |
| const { modelGroupMap, groupMap } = azureConfig; |
| const { |
| azureOptions, |
| baseURL, |
| headers = {}, |
| serverless: _serverless, |
| } = mapModelToAzureConfig({ |
| modelName, |
| modelGroupMap, |
| groupMap, |
| }); |
| serverless = _serverless; |
|
|
| clientOptions.reverseProxyUrl = baseURL ?? clientOptions.reverseProxyUrl; |
| clientOptions.headers = resolveHeaders({ |
| headers: { ...headers, ...(clientOptions.headers ?? {}) }, |
| user: req.user, |
| }); |
|
|
| clientOptions.titleConvo = azureConfig.titleConvo; |
| clientOptions.titleModel = azureConfig.titleModel; |
|
|
| const azureRate = modelName.includes('gpt-4') ? 30 : 17; |
| clientOptions.streamRate = azureConfig.streamRate ?? azureRate; |
|
|
| clientOptions.titleMethod = azureConfig.titleMethod ?? 'completion'; |
|
|
| const groupName = modelGroupMap[modelName].group; |
| clientOptions.addParams = azureConfig.groupMap[groupName].addParams; |
| clientOptions.dropParams = azureConfig.groupMap[groupName].dropParams; |
| clientOptions.forcePrompt = azureConfig.groupMap[groupName].forcePrompt; |
|
|
| apiKey = azureOptions.azureOpenAIApiKey; |
| clientOptions.azure = !serverless && azureOptions; |
| if (serverless === true) { |
| clientOptions.defaultQuery = azureOptions.azureOpenAIApiVersion |
| ? { 'api-version': azureOptions.azureOpenAIApiVersion } |
| : undefined; |
| clientOptions.headers['api-key'] = apiKey; |
| } |
| } else if (isAzureOpenAI) { |
| clientOptions.azure = userProvidesKey ? JSON.parse(userValues.apiKey) : getAzureCredentials(); |
| apiKey = clientOptions.azure.azureOpenAIApiKey; |
| } |
|
|
| |
| const openAIConfig = appConfig.endpoints?.[EModelEndpoint.openAI]; |
|
|
| if (!isAzureOpenAI && openAIConfig) { |
| clientOptions.streamRate = openAIConfig.streamRate; |
| clientOptions.titleModel = openAIConfig.titleModel; |
| } |
|
|
| const allConfig = appConfig.endpoints?.all; |
| if (allConfig) { |
| clientOptions.streamRate = allConfig.streamRate; |
| } |
|
|
| if (userProvidesKey & !apiKey) { |
| throw new Error( |
| JSON.stringify({ |
| type: ErrorTypes.NO_USER_KEY, |
| }), |
| ); |
| } |
|
|
| if (!apiKey) { |
| throw new Error(`${endpoint} API Key not provided.`); |
| } |
|
|
| if (optionsOnly) { |
| const modelOptions = endpointOption?.model_parameters ?? {}; |
| modelOptions.model = modelName; |
| clientOptions = Object.assign({ modelOptions }, clientOptions); |
| clientOptions.modelOptions.user = req.user.id; |
| const options = getOpenAIConfig(apiKey, clientOptions, endpoint); |
| if (options != null && serverless === true) { |
| options.useLegacyContent = true; |
| } |
| const streamRate = clientOptions.streamRate; |
| if (!streamRate) { |
| return options; |
| } |
| options.llmConfig._lc_stream_delay = streamRate; |
| return options; |
| } |
|
|
| const client = new OpenAIClient(apiKey, Object.assign({ req, res }, clientOptions)); |
| return { |
| client, |
| openAIApiKey: apiKey, |
| }; |
| }; |
|
|
| module.exports = initializeClient; |
|
|