Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Azimuth UI improvements #5

Merged
merged 7 commits into from
Feb 1, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ The following is a non-exhaustive list of models which have been tested with thi
- [Mistral 7B Instruct v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
<!-- - [AWQ Quantized Mixtral 8x7B Instruct v0.1](https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ) (Not producing output properly) -->

Due to the combination of [components](##Components) used in this app, some HuggingFace models may not work as expected (usually due to the way in which LangChain formats the prompt messages). Any errors when using new model will appear in the pod logs for either the web-app deployment or the backend API deployment.
Due to the combination of [components](##Components) used in this app, some HuggingFace models may not work as expected (usually due to the way in which LangChain formats the prompt messages). Any errors when using new model will appear in the logs for either the web-app pod or the backend API pod. Please open an issue if you would like explicit support for a specific model which is not in the above list.


## Components
Expand Down
1 change: 1 addition & 0 deletions chart/azimuth-ui.schema.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
controls:
/huggingface/model:
type: TextControl
required: true
/huggingface/token:
type: TextControl
secret: true
Expand Down
28 changes: 18 additions & 10 deletions chart/values.schema.json
Original file line number Diff line number Diff line change
Expand Up @@ -8,12 +8,12 @@
"model": {
"type": "string",
"title": "Model",
"description": "The HuggingFace model to deploy (Hint: For a simple, lightweight demo try ise-uiuc/Magicoder-S-DS-6.7B)"
"description": "The [HuggingFace model](https://huggingface.co/models) to deploy (see [here](https://github.com/stackhpc/azimuth-llm?tab=readme-ov-file#tested-models) for a list of tested models)."
},
"token": {
"type": "string",
"title": "Access Token",
"description": "The HuggingFace access token to use for installing gated models.",
"description": "A HuggingFace [access token](https://huggingface.co/docs/hub/security-tokens). Only required for [gated models](https://huggingface.co/docs/hub/en/models-gated (e.g. Llama 2).",
"default": ""
}
},
Expand All @@ -28,42 +28,50 @@
"model_name": {
"type": "string",
"title": "Model Name",
"description": "Model name supplied to OpenAI client in frontend web app. Should match huggingface.model above."
"description": "Model name supplied to the OpenAI client in frontend web app. Should match huggingface.model above.",
"default": "mistralai/Mistral-7B-Instruct-v0.2"
},
"model_instruction": {
"type": "string",
"title": "Instruction",
"description": "The initial model prompt (i.e. the hidden instructions) to use when generating responses."
"description": "The initial model prompt (i.e. the hidden instructions) to use when generating responses.",
"default": "You are a helpful AI assistant. Please respond appropriately."
},
"page_title": {
"type": "string",
"title": "Page Title",
"description": "The title to use for the chat interface."
"description": "The title to use for the chat interface.",
"default": "Large Language Model"
},
"llm_max_tokens": {
"type": "number",
"title": "Max Tokens",
"description": "The maximum number of new [tokens](https://platform.openai.com/docs/api-reference/chat/create#chat-create-max_tokens) to generate for each LLM responses."
"description": "The maximum number of new [tokens](https://platform.openai.com/docs/api-reference/chat/create#chat-create-max_tokens) to generate for each LLM responses.",
"default": 1000
},
"llm_temperature": {
"type": "number",
"title": "LLM Temperature",
"description": "The '[temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature)' value to use when generating LLM responses."
"description": "The [temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature) value to use when generating LLM responses.",
"default": 1
},
"llm_top_p": {
"type": "number",
"title": "LLM Top P",
"description": "The [top p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) value to use when generating LLM responses."
"description": "The [top p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) value to use when generating LLM responses.",
"default": 1
},
"llm_presence_penalty": {
"type": "number",
"title": "LLM Presence Penalty",
"description": "The [presence penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-presence_penalty) to use when generating LLM responses."
"description": "The [presence penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-presence_penalty) to use when generating LLM responses.",
"default": 0
},
"llm_frequency_penalty": {
"type": "number",
"title": "LLM Frequency Penalty",
"description": "The [frequency_penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-frequency_penalty) to use when generating LLM responses."
"description": "The [frequency_penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-frequency_penalty) to use when generating LLM responses.",
"default": 0
}

},
Expand Down
Loading