Skip to content

Error When Accessing Azure Serverless Models #4771

Discussion options

You must be logged in to vote

I'm updating the docs, but the update is now live:

Here is an example configuration for Meta-Llama-3.1-8B-Instruct:

endpoints:
  azureOpenAI:
    groups:
    - group: "serverless-example"
      apiKey: "${LLAMA318B_API_KEY}"
      baseURL: "https://example.services.ai.azure.com/models/"
      version: "2024-05-01-preview" # Optional: specify API version
      serverless: true
      models:
        # Must match the deployment name of the model
        Meta-Llama-3.1-8B-Instruct: true

Notes:

  • Azure AI Foundry models now provision endpoints under /models/chat/completions?api-version=version for serverless inference.
    • The baseURL field should be set to the root of the endpoint, without anythi…

Replies: 2 comments 2 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
2 replies
@MasterOfWebM
Comment options

@danny-avila
Comment options

Answer selected by MasterOfWebM
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #4767 on November 21, 2024 14:56.