Skip to content
This repository has been archived by the owner on Oct 7, 2024. It is now read-only.

Support for Dynamic Quota in Chat-Models #91

Open
1 task done
g4challenge opened this issue Aug 6, 2024 · 2 comments
Open
1 task done

Support for Dynamic Quota in Chat-Models #91

g4challenge opened this issue Aug 6, 2024 · 2 comments

Comments

@g4challenge
Copy link

Is there an existing issue for this?

  • I have searched the existing issues

Description

I would need to enable dynamic quota for the default gpt models

New or Affected Resource(s)/Data Source(s)

openai_deployment

Potential Terraform Configuration

eployment = {                                # TODO make sure to update corresponding litellm config
    "chat_model" = {
      name          = "gpt-4o"
      model_format  = "OpenAI"
      model_name    = "gpt-4o"
      model_version = "2024-05-13"
      scale_type    = "Standard"
      dynamic_quota_enabled = true
      #capacity      = 120
    },

References

https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/dynamic-quota

@zioproto
Copy link
Collaborator

zioproto commented Aug 6, 2024

Hello @g4challenge,

TL;DR waiting on Terraform provider feature

to implement this feature in the module we need first the azurerm provider to support this feature.

I see there is a open issue hashicorp/terraform-provider-azurerm#23988 and an existing PR hashicorp/terraform-provider-azurerm#25401

We have to wait for the feature to be merged.

thanks

@zioproto
Copy link
Collaborator

zioproto commented Aug 6, 2024

Hello @g4challenge

I also noticed that the feature is in preview. We will add the feature to the module only when it is promoted to GA.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
Development

No branches or pull requests

2 participants