diff --git a/packages/unillm-node/package.json b/packages/unillm-node/package.json index 228d074..4564efc 100644 --- a/packages/unillm-node/package.json +++ b/packages/unillm-node/package.json @@ -1,6 +1,6 @@ { "name": "unillm", - "version": "0.1.0-alpha.2", + "version": "0.1.0", "main": "./index.ts", "type": "commonjs", "license": "MIT", diff --git a/packages/unillm-node/providers/azure-openai.ts b/packages/unillm-node/providers/azure-openai.ts index abd689f..56cf4bd 100644 --- a/packages/unillm-node/providers/azure-openai.ts +++ b/packages/unillm-node/providers/azure-openai.ts @@ -92,10 +92,12 @@ export class AzureOpenAIProvider extends BaseProvider { } async createChatCompletionStreaming( - model: ModelTypes[Providers.AzureOpenAI], + _model: ModelTypes[Providers.AzureOpenAI], params: UnifiedCreateChatCompletionParamsStreaming, ): Promise { + const [, model] = _model.split("/"); const { baseParams } = this.processUnifiedParamsToAzureOpenAIFormat(params); + const originalStreamResponse = this.client.listChatCompletions( model, params.messages,