|
import { BaseProvider } from '~/lib/modules/llm/base-provider'; |
|
import type { ModelInfo } from '~/lib/modules/llm/types'; |
|
import type { IProviderSetting } from '~/types/model'; |
|
import { createOpenAI } from '@ai-sdk/openai'; |
|
import type { LanguageModelV1 } from 'ai'; |
|
import { logger } from '~/utils/logger'; |
|
|
|
export default class LMStudioProvider extends BaseProvider { |
|
name = 'LMStudio'; |
|
getApiKeyLink = 'https://lmstudio.ai/'; |
|
labelForGetApiKey = 'Get LMStudio'; |
|
icon = 'i-ph:cloud-arrow-down'; |
|
|
|
config = { |
|
baseUrlKey: 'LMSTUDIO_API_BASE_URL', |
|
baseUrl: 'http://localhost:1234/', |
|
}; |
|
|
|
staticModels: ModelInfo[] = []; |
|
|
|
async getDynamicModels( |
|
apiKeys?: Record<string, string>, |
|
settings?: IProviderSetting, |
|
serverEnv: Record<string, string> = {}, |
|
): Promise<ModelInfo[]> { |
|
let { baseUrl } = this.getProviderBaseUrlAndKey({ |
|
apiKeys, |
|
providerSettings: settings, |
|
serverEnv, |
|
defaultBaseUrlKey: 'LMSTUDIO_API_BASE_URL', |
|
defaultApiTokenKey: '', |
|
}); |
|
|
|
if (!baseUrl) { |
|
throw new Error('No baseUrl found for LMStudio provider'); |
|
} |
|
|
|
if (typeof window === 'undefined') { |
|
|
|
|
|
|
|
|
|
const isDocker = process?.env?.RUNNING_IN_DOCKER === 'true' || serverEnv?.RUNNING_IN_DOCKER === 'true'; |
|
|
|
baseUrl = isDocker ? baseUrl.replace('localhost', 'host.docker.internal') : baseUrl; |
|
baseUrl = isDocker ? baseUrl.replace('127.0.0.1', 'host.docker.internal') : baseUrl; |
|
} |
|
|
|
const response = await fetch(`${baseUrl}/v1/models`); |
|
const data = (await response.json()) as { data: Array<{ id: string }> }; |
|
|
|
return data.data.map((model) => ({ |
|
name: model.id, |
|
label: model.id, |
|
provider: this.name, |
|
maxTokenAllowed: 8000, |
|
})); |
|
} |
|
getModelInstance: (options: { |
|
model: string; |
|
serverEnv?: Env; |
|
apiKeys?: Record<string, string>; |
|
providerSettings?: Record<string, IProviderSetting>; |
|
}) => LanguageModelV1 = (options) => { |
|
const { apiKeys, providerSettings, serverEnv, model } = options; |
|
let { baseUrl } = this.getProviderBaseUrlAndKey({ |
|
apiKeys, |
|
providerSettings: providerSettings?.[this.name], |
|
serverEnv: serverEnv as any, |
|
defaultBaseUrlKey: 'LMSTUDIO_API_BASE_URL', |
|
defaultApiTokenKey: '', |
|
}); |
|
|
|
if (!baseUrl) { |
|
throw new Error('No baseUrl found for LMStudio provider'); |
|
} |
|
|
|
const isDocker = process.env.RUNNING_IN_DOCKER === 'true' || serverEnv?.RUNNING_IN_DOCKER === 'true'; |
|
|
|
if (typeof window === 'undefined') { |
|
baseUrl = isDocker ? baseUrl.replace('localhost', 'host.docker.internal') : baseUrl; |
|
baseUrl = isDocker ? baseUrl.replace('127.0.0.1', 'host.docker.internal') : baseUrl; |
|
} |
|
|
|
logger.debug('LMStudio Base Url used: ', baseUrl); |
|
|
|
const lmstudio = createOpenAI({ |
|
baseURL: `${baseUrl}/v1`, |
|
apiKey: '', |
|
}); |
|
|
|
return lmstudio(model); |
|
}; |
|
} |
|
|