Skip to content

Conversation

@pauldotyu
Copy link
Contributor

Purpose

  • This enhancement is to add support for LLMs running locally in the cluster (i.e., models and inferencing endpoints deployed by KAITO).
  • To use local LLMs, you will need to set the USE_LOCAL_LLM environment variable in the ai-service.yaml manifest to "True" and AI_ENDPOINT to the inferencing service that is running in the cluster

Does this introduce a breaking change?

[ ] Yes
[x] No

Pull Request Type

What kind of change does this Pull Request introduce?

[ ] Bugfix
[x] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ ] Other... Please describe:

How to Test

  • Ensure this doc still works https://learn.microsoft.com/en-us/azure/aks/open-ai-quickstart?tabs=aoai
  • To test local LLM, deploy KAITO to AKS cluster and update ai-service.yaml manifest to set USE_LOCAL_LLM to "True" and AI_ENDPOINT to the inferencing service that KAITO configures. Also be sure to unset the OpenAI specific environment variables (e.g., USE_AZURE_OPENAI, AZURE_OPENAI_ENDPOINT, AZURE_OPENAI_DEPLOYMENT_NAME, OPENAI_API_KEY, and OPENAI_ORG_ID in the ai-service.yaml manifest)

@pauldotyu pauldotyu requested a review from chzbrgr71 November 15, 2023 21:21
Copy link
Collaborator

@chzbrgr71 chzbrgr71 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Completed some basic local testing. Will confirm again once new image is created

@chzbrgr71 chzbrgr71 merged commit fa0cb75 into Azure-Samples:main Nov 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants